[ 488.169687] env[62521]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62521) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.170028] env[62521]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62521) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.170132] env[62521]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62521) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.170418] env[62521]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 488.264265] env[62521]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62521) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 488.274754] env[62521]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62521) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 488.873603] env[62521]: INFO nova.virt.driver [None req-3bfc8bb9-9766-49a3-a47c-e5d5cfd5e96f None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 488.942955] env[62521]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 488.943148] env[62521]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 488.943232] env[62521]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62521) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 492.051352] env[62521]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-0789dabd-8ba3-4195-a072-be6e6f074f3e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.067095] env[62521]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62521) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 492.067253] env[62521]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-4e4e56d3-a03e-4a59-b03d-fb0dc732b9c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.101103] env[62521]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 00d39. [ 492.101262] env[62521]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.158s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 492.101847] env[62521]: INFO nova.virt.vmwareapi.driver [None req-3bfc8bb9-9766-49a3-a47c-e5d5cfd5e96f None None] VMware vCenter version: 7.0.3 [ 492.105253] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42049de-e308-403a-819e-cdb08fdad7c6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.126154] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130068d3-8122-4733-a723-a0b5026f7956 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.131893] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34249f02-eaa3-4680-82d4-7933500d4317 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.138375] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa8f2f8-7e36-410d-8e9a-787517f1789d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.151182] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63ab1fb-156e-479b-b060-51b4c8afb50c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.156928] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0876e681-6a56-4110-9d6f-6900d0c27bab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.186822] env[62521]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-b18d733f-4b45-42d0-ac6f-399d27d4a9ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.192188] env[62521]: DEBUG nova.virt.vmwareapi.driver [None req-3bfc8bb9-9766-49a3-a47c-e5d5cfd5e96f None None] Extension org.openstack.compute already exists. {{(pid=62521) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 492.194873] env[62521]: INFO nova.compute.provider_config [None req-3bfc8bb9-9766-49a3-a47c-e5d5cfd5e96f None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 492.698680] env[62521]: DEBUG nova.context [None req-3bfc8bb9-9766-49a3-a47c-e5d5cfd5e96f None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),8c5ff770-5d01-44bc-9fa2-d5e0a983b355(cell1) {{(pid=62521) load_cells /opt/stack/nova/nova/context.py:464}} [ 492.700886] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.701140] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 492.701847] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 492.702302] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Acquiring lock "8c5ff770-5d01-44bc-9fa2-d5e0a983b355" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.702563] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Lock "8c5ff770-5d01-44bc-9fa2-d5e0a983b355" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 492.703617] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Lock "8c5ff770-5d01-44bc-9fa2-d5e0a983b355" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 492.723619] env[62521]: INFO dbcounter [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Registered counter for database nova_cell0 [ 492.731616] env[62521]: INFO dbcounter [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Registered counter for database nova_cell1 [ 492.735058] env[62521]: DEBUG oslo_db.sqlalchemy.engines [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62521) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 492.735670] env[62521]: DEBUG oslo_db.sqlalchemy.engines [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62521) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 492.740014] env[62521]: ERROR nova.db.main.api [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 492.740014] env[62521]: result = function(*args, **kwargs) [ 492.740014] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 492.740014] env[62521]: return func(*args, **kwargs) [ 492.740014] env[62521]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 492.740014] env[62521]: result = fn(*args, **kwargs) [ 492.740014] env[62521]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 492.740014] env[62521]: return f(*args, **kwargs) [ 492.740014] env[62521]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 492.740014] env[62521]: return db.service_get_minimum_version(context, binaries) [ 492.740014] env[62521]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 492.740014] env[62521]: _check_db_access() [ 492.740014] env[62521]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 492.740014] env[62521]: stacktrace = ''.join(traceback.format_stack()) [ 492.740014] env[62521]: [ 492.740974] env[62521]: ERROR nova.db.main.api [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 492.740974] env[62521]: result = function(*args, **kwargs) [ 492.740974] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 492.740974] env[62521]: return func(*args, **kwargs) [ 492.740974] env[62521]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 492.740974] env[62521]: result = fn(*args, **kwargs) [ 492.740974] env[62521]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 492.740974] env[62521]: return f(*args, **kwargs) [ 492.740974] env[62521]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 492.740974] env[62521]: return db.service_get_minimum_version(context, binaries) [ 492.740974] env[62521]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 492.740974] env[62521]: _check_db_access() [ 492.740974] env[62521]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 492.740974] env[62521]: stacktrace = ''.join(traceback.format_stack()) [ 492.740974] env[62521]: [ 492.741711] env[62521]: WARNING nova.objects.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 492.741711] env[62521]: WARNING nova.objects.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Failed to get minimum service version for cell 8c5ff770-5d01-44bc-9fa2-d5e0a983b355 [ 492.741935] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Acquiring lock "singleton_lock" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 492.742113] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Acquired lock "singleton_lock" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 492.742361] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Releasing lock "singleton_lock" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 492.742690] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Full set of CONF: {{(pid=62521) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 492.742834] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ******************************************************************************** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 492.742963] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Configuration options gathered from: {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 492.743120] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 492.743339] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 492.743480] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ================================================================================ {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 492.743691] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] allow_resize_to_same_host = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.743861] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] arq_binding_timeout = 300 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.743993] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] backdoor_port = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.744139] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] backdoor_socket = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.744325] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] block_device_allocate_retries = 60 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.744547] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] block_device_allocate_retries_interval = 3 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.744726] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cert = self.pem {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.744892] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.745071] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute_monitors = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.745240] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] config_dir = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.745409] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] config_drive_format = iso9660 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.745543] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.745708] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] config_source = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.745874] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] console_host = devstack {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.746049] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] control_exchange = nova {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.746212] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cpu_allocation_ratio = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.746369] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] daemon = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.746536] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] debug = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.746691] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] default_access_ip_network_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.746853] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] default_availability_zone = nova {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.747011] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] default_ephemeral_format = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.747173] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] default_green_pool_size = 1000 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.747409] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.747566] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] default_schedule_zone = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.747720] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] disk_allocation_ratio = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.747878] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] enable_new_services = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.748063] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] enabled_apis = ['osapi_compute'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.748230] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] enabled_ssl_apis = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.748390] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] flat_injected = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.748549] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] force_config_drive = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.748706] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] force_raw_images = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.748873] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] graceful_shutdown_timeout = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.749042] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] heal_instance_info_cache_interval = 60 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.749258] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] host = cpu-1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.749432] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.749596] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.749768] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.749972] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.750154] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] instance_build_timeout = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.750316] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] instance_delete_interval = 300 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.750484] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] instance_format = [instance: %(uuid)s] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.750651] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] instance_name_template = instance-%08x {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.750812] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] instance_usage_audit = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.750981] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] instance_usage_audit_period = month {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.751160] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.751348] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.751584] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] internal_service_availability_zone = internal {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.751767] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] key = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.751934] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] live_migration_retry_count = 30 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.752121] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] log_color = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.752289] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] log_config_append = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.752459] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.752619] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] log_dir = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.752775] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] log_file = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.752905] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] log_options = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.753076] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] log_rotate_interval = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.753247] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] log_rotate_interval_type = days {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.753445] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] log_rotation_type = none {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.753585] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.753714] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.753885] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.754063] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.754194] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.754363] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] long_rpc_timeout = 1800 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.754545] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] max_concurrent_builds = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.754709] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] max_concurrent_live_migrations = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.754869] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] max_concurrent_snapshots = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.755038] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] max_local_block_devices = 3 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.755202] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] max_logfile_count = 30 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.755361] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] max_logfile_size_mb = 200 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.755522] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] maximum_instance_delete_attempts = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.755690] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] metadata_listen = 0.0.0.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.755860] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] metadata_listen_port = 8775 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.756036] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] metadata_workers = 2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.756203] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] migrate_max_retries = -1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.756370] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] mkisofs_cmd = genisoimage {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.756576] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.756707] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] my_ip = 10.180.1.21 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.756867] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] network_allocate_retries = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.757053] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.757223] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.757384] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] osapi_compute_listen_port = 8774 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.757552] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] osapi_compute_unique_server_name_scope = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.757716] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] osapi_compute_workers = 2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.757876] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] password_length = 12 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.758043] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] periodic_enable = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.758205] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] periodic_fuzzy_delay = 60 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.758372] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] pointer_model = usbtablet {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.758539] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] preallocate_images = none {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.758696] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] publish_errors = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.758823] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] pybasedir = /opt/stack/nova {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.758978] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ram_allocation_ratio = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.759151] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] rate_limit_burst = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.759315] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] rate_limit_except_level = CRITICAL {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.759474] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] rate_limit_interval = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.759629] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] reboot_timeout = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.759785] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] reclaim_instance_interval = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.759938] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] record = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.760114] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] reimage_timeout_per_gb = 60 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.760278] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] report_interval = 120 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.760438] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] rescue_timeout = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.760594] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] reserved_host_cpus = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.760749] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] reserved_host_disk_mb = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.760903] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] reserved_host_memory_mb = 512 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.761068] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] reserved_huge_pages = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.761227] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] resize_confirm_window = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.761407] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] resize_fs_using_block_device = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.761571] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] resume_guests_state_on_host_boot = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.761736] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.761895] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] rpc_response_timeout = 60 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.762070] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] run_external_periodic_tasks = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.762241] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] running_deleted_instance_action = reap {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.762400] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.762559] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] running_deleted_instance_timeout = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.762713] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler_instance_sync_interval = 120 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.762915] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_down_time = 720 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.763170] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] servicegroup_driver = db {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.763376] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] shell_completion = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.763547] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] shelved_offload_time = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.763710] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] shelved_poll_interval = 3600 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.763877] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] shutdown_timeout = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.764048] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] source_is_ipv6 = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.764213] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ssl_only = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.764459] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.764629] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] sync_power_state_interval = 600 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.764792] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] sync_power_state_pool_size = 1000 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.764956] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] syslog_log_facility = LOG_USER {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.765125] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] tempdir = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.765285] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] timeout_nbd = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.765454] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] transport_url = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.765616] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] update_resources_interval = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.765773] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] use_cow_images = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.765930] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] use_eventlog = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.766101] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] use_journal = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.766261] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] use_json = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.766419] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] use_rootwrap_daemon = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.766577] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] use_stderr = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.766732] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] use_syslog = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.766886] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vcpu_pin_set = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.767062] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plugging_is_fatal = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.767230] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plugging_timeout = 300 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.767396] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] virt_mkfs = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.767558] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] volume_usage_poll_interval = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.767716] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] watch_log_file = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.767883] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] web = /usr/share/spice-html5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.768081] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.768252] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.768416] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.768587] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_concurrency.disable_process_locking = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.769164] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.769320] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.769497] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.769673] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.769845] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.770020] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.770209] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.auth_strategy = keystone {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.770379] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.compute_link_prefix = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.770557] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.770732] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.dhcp_domain = novalocal {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.770968] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.enable_instance_password = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.771171] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.glance_link_prefix = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.771357] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.771537] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.771706] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.instance_list_per_project_cells = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.771873] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.list_records_by_skipping_down_cells = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.772055] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.local_metadata_per_cell = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.772233] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.max_limit = 1000 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.772402] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.metadata_cache_expiration = 15 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.772578] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.neutron_default_tenant_id = default {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.772748] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.response_validation = warn {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.772921] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.use_neutron_default_nets = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.773102] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.773273] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.773473] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.773657] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.773833] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.vendordata_dynamic_targets = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.773996] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.vendordata_jsonfile_path = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.774194] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.774423] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.backend = dogpile.cache.memcached {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.774605] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.backend_argument = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.774779] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.config_prefix = cache.oslo {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.774948] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.dead_timeout = 60.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.775125] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.debug_cache_backend = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.775289] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.enable_retry_client = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.775450] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.enable_socket_keepalive = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.775620] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.enabled = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.775782] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.enforce_fips_mode = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.775944] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.expiration_time = 600 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.776116] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.hashclient_retry_attempts = 2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.776282] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.776444] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.memcache_dead_retry = 300 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.776599] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.memcache_password = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.776761] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.776922] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.777093] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.memcache_pool_maxsize = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.777257] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.777419] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.memcache_sasl_enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.777596] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.777761] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.777920] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.memcache_username = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.778094] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.proxies = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.778261] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.redis_db = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.778419] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.redis_password = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.778588] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.778764] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.778930] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.redis_server = localhost:6379 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.779107] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.redis_socket_timeout = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.779269] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.redis_username = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.779432] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.retry_attempts = 2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.779597] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.retry_delay = 0.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.779757] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.socket_keepalive_count = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.779917] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.socket_keepalive_idle = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.780089] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.socket_keepalive_interval = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.780250] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.tls_allowed_ciphers = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.780408] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.tls_cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.780566] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.tls_certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.780728] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.tls_enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.780882] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cache.tls_keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.781060] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.auth_section = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.781234] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.auth_type = password {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.781419] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.781600] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.781758] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.781919] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.782089] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.cross_az_attach = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.782251] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.debug = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.782408] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.endpoint_template = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.782572] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.http_retries = 3 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.782732] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.782886] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.783070] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.os_region_name = RegionOne {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.783239] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.783425] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cinder.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.783603] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.783767] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.cpu_dedicated_set = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.783926] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.cpu_shared_set = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.784103] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.image_type_exclude_list = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.784278] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.784464] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.784652] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.784787] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.784956] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.785132] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.resource_provider_association_refresh = 300 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.785297] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.785463] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.shutdown_retry_interval = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.785640] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.785816] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] conductor.workers = 2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.785995] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] console.allowed_origins = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.786169] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] console.ssl_ciphers = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.786339] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] console.ssl_minimum_version = default {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.786511] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] consoleauth.enforce_session_timeout = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.786676] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] consoleauth.token_ttl = 600 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.786843] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.786999] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.787177] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.787338] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.connect_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.787496] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.connect_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.787653] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.endpoint_override = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.787814] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.787969] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.788142] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.max_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.788298] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.min_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.788458] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.region_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.788615] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.retriable_status_codes = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.788770] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.service_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.788937] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.service_type = accelerator {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.789108] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.789267] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.status_code_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.789424] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.status_code_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.789581] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.789762] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.789923] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] cyborg.version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.790119] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.backend = sqlalchemy {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.790290] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.connection = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.790457] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.connection_debug = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.790624] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.connection_parameters = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.790787] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.connection_recycle_time = 3600 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.790949] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.connection_trace = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.791123] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.db_inc_retry_interval = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.791298] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.db_max_retries = 20 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.791486] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.db_max_retry_interval = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.791654] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.db_retry_interval = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.791817] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.max_overflow = 50 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.791979] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.max_pool_size = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.792153] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.max_retries = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.792331] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.792481] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.mysql_wsrep_sync_wait = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.792638] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.pool_timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.792798] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.retry_interval = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.792954] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.slave_connection = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.793128] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.sqlite_synchronous = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.793297] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] database.use_db_reconnect = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.793501] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.backend = sqlalchemy {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.793674] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.connection = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.793838] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.connection_debug = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.794022] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.connection_parameters = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.794186] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.connection_recycle_time = 3600 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.794373] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.connection_trace = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.794542] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.db_inc_retry_interval = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.794712] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.db_max_retries = 20 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.794867] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.db_max_retry_interval = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.795044] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.db_retry_interval = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.795212] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.max_overflow = 50 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.795374] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.max_pool_size = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.795535] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.max_retries = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.795703] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.795862] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.796028] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.pool_timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.796196] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.retry_interval = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.796355] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.slave_connection = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.796516] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] api_database.sqlite_synchronous = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.796693] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] devices.enabled_mdev_types = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.796870] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.797050] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.797217] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ephemeral_storage_encryption.enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.797382] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.797555] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.api_servers = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.797717] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.797880] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.798056] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.798221] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.connect_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.798381] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.connect_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.798546] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.debug = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.798710] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.default_trusted_certificate_ids = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.798870] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.enable_certificate_validation = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.799044] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.enable_rbd_download = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.799207] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.endpoint_override = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.799373] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.799537] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.799695] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.max_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.799852] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.min_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.800023] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.num_retries = 3 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.800197] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.rbd_ceph_conf = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.800360] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.rbd_connect_timeout = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.800529] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.rbd_pool = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.800694] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.rbd_user = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.800853] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.region_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.801022] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.retriable_status_codes = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.801183] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.service_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.801372] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.service_type = image {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.801550] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.801718] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.status_code_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.801877] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.status_code_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.802049] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.802234] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.802400] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.verify_glance_signatures = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.802562] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] glance.version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.802729] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] guestfs.debug = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.802894] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] mks.enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.803258] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.803478] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] image_cache.manager_interval = 2400 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.803657] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] image_cache.precache_concurrency = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.803828] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] image_cache.remove_unused_base_images = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.804006] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.804186] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.804365] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] image_cache.subdirectory_name = _base {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.804543] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.api_max_retries = 60 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.804707] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.api_retry_interval = 2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.804908] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.auth_section = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.805039] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.auth_type = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.805205] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.805363] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.805529] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.805694] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.conductor_group = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.805854] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.connect_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.806050] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.connect_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.806182] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.endpoint_override = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.806345] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.806504] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.806661] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.max_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.806816] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.min_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.806981] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.peer_list = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.807151] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.region_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.807310] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.retriable_status_codes = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.807476] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.serial_console_state_timeout = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.807635] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.service_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.807803] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.service_type = baremetal {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.807963] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.shard = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.808146] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.808307] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.status_code_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.808468] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.status_code_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.808626] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.808805] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.808965] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ironic.version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.809158] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.809331] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] key_manager.fixed_key = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.809553] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.809671] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.barbican_api_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.809829] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.barbican_endpoint = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.809997] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.barbican_endpoint_type = public {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.810169] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.barbican_region_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.810328] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.810491] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.810651] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.810809] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.810965] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.811140] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.number_of_retries = 60 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.811316] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.retry_delay = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.811495] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.send_service_user_token = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.811659] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.811818] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.811979] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.verify_ssl = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.812150] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican.verify_ssl_path = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.812315] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican_service_user.auth_section = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.812478] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican_service_user.auth_type = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.812636] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican_service_user.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.812791] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican_service_user.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.812952] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican_service_user.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.813125] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican_service_user.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.813315] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican_service_user.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.813531] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican_service_user.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.813716] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] barbican_service_user.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.813888] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.approle_role_id = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.814064] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.approle_secret_id = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.814252] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.kv_mountpoint = secret {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.814434] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.kv_path = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.814610] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.kv_version = 2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.814772] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.namespace = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.814930] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.root_token_id = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.815104] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.ssl_ca_crt_file = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.815273] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.timeout = 60.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.815435] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.use_ssl = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.815606] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.815778] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.auth_section = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.815941] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.auth_type = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.816112] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.816274] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.816440] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.816600] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.connect_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.816759] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.connect_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.816917] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.endpoint_override = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.817092] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.817255] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.817414] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.max_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.817574] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.min_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.817733] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.region_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.817889] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.retriable_status_codes = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.818056] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.service_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.818230] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.service_type = identity {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.818446] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.818620] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.status_code_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.818784] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.status_code_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.818962] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.819176] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.819344] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] keystone.version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.819547] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.connection_uri = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.819713] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.cpu_mode = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.819879] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.820059] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.cpu_models = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.820236] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.cpu_power_governor_high = performance {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.820403] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.820569] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.cpu_power_management = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.820805] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.820987] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.device_detach_attempts = 8 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.821182] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.device_detach_timeout = 20 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.821422] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.disk_cachemodes = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.821629] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.disk_prefix = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.821802] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.enabled_perf_events = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.821999] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.file_backed_memory = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.822193] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.gid_maps = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.822370] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.hw_disk_discard = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.822548] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.hw_machine_type = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.822726] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.images_rbd_ceph_conf = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.822896] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.823073] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.823245] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.images_rbd_glance_store_name = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.823443] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.images_rbd_pool = rbd {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.823626] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.images_type = default {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.823789] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.images_volume_group = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.823952] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.inject_key = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.824128] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.inject_partition = -2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.824292] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.inject_password = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.824462] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.iscsi_iface = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.824682] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.iser_use_multipath = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.824860] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.825035] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.825206] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_downtime = 500 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.825372] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.825537] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.825701] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_inbound_addr = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.825863] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.826040] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.826207] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_scheme = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.826381] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_timeout_action = abort {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.826549] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_tunnelled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.826710] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_uri = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.826874] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.live_migration_with_native_tls = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.827056] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.max_queues = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.827244] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.827503] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.827674] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.nfs_mount_options = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.827970] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.828158] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.828328] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.828492] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.828656] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.828820] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.num_pcie_ports = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.828986] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.829168] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.pmem_namespaces = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.829330] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.quobyte_client_cfg = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.829618] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.829794] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.829959] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.830137] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.830300] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.rbd_secret_uuid = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.830460] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.rbd_user = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.830623] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.830793] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.830954] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.rescue_image_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.831132] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.rescue_kernel_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.831307] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.rescue_ramdisk_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.831504] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.831671] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.rx_queue_size = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.831842] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.smbfs_mount_options = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.832129] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.832304] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.snapshot_compression = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.832469] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.snapshot_image_format = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.832685] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.832853] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.sparse_logical_volumes = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.833031] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.swtpm_enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.833205] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.swtpm_group = tss {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.833399] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.swtpm_user = tss {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.833579] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.sysinfo_serial = unique {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.833740] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.tb_cache_size = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.833898] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.tx_queue_size = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.834076] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.uid_maps = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.834244] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.use_virtio_for_bridges = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.834413] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.virt_type = kvm {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.834583] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.volume_clear = zero {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.834745] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.volume_clear_size = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.834909] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.volume_use_multipath = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.835092] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.vzstorage_cache_path = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.835268] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.835435] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.835606] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.835771] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.836058] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.836240] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.vzstorage_mount_user = stack {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.836411] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.836592] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.auth_section = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.836765] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.auth_type = password {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.836928] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.837100] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.837268] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.837429] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.connect_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.837592] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.connect_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.837763] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.default_floating_pool = public {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.837923] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.endpoint_override = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.838099] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.extension_sync_interval = 600 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.838265] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.http_retries = 3 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.838428] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.838591] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.838751] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.max_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.838920] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.839091] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.min_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.839262] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.ovs_bridge = br-int {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.839427] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.physnets = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.839597] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.region_name = RegionOne {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.839758] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.retriable_status_codes = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.839929] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.service_metadata_proxy = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.840099] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.service_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.840270] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.service_type = network {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.840435] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.840598] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.status_code_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.840757] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.status_code_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.840917] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.841109] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.841282] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] neutron.version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.841488] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] notifications.bdms_in_notifications = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.841674] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] notifications.default_level = INFO {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.841851] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] notifications.notification_format = unversioned {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.842027] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] notifications.notify_on_state_change = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.842212] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.842389] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] pci.alias = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.842565] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] pci.device_spec = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.842731] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] pci.report_in_placement = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.842904] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.auth_section = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.843091] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.auth_type = password {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.843267] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.843455] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.843622] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.843787] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.843947] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.connect_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.844119] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.connect_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.844282] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.default_domain_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.844442] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.default_domain_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.844599] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.domain_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.844754] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.domain_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.844909] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.endpoint_override = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.845080] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.845241] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.845403] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.max_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.845560] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.min_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.845727] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.password = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.845885] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.project_domain_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.846061] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.project_domain_name = Default {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.846236] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.project_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.846408] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.project_name = service {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.846578] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.region_name = RegionOne {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.846740] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.retriable_status_codes = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.846900] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.service_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.847081] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.service_type = placement {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.847248] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.847409] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.status_code_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.847572] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.status_code_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.847733] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.system_scope = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.847891] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.848058] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.trust_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.848219] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.user_domain_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.848387] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.user_domain_name = Default {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.848547] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.user_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.848718] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.username = nova {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.848897] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.849069] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] placement.version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.849251] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.cores = 20 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.849417] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.count_usage_from_placement = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.849590] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.849765] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.injected_file_content_bytes = 10240 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.849938] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.injected_file_path_length = 255 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.850106] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.injected_files = 5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.850275] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.instances = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.850443] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.key_pairs = 100 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.850612] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.metadata_items = 128 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.850775] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.ram = 51200 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.850936] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.recheck_quota = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.851117] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.server_group_members = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.851296] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] quota.server_groups = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.851504] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.851675] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.851840] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler.image_metadata_prefilter = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.852007] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.852183] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler.max_attempts = 3 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.852346] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler.max_placement_results = 1000 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.852514] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.852676] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.852837] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.853014] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] scheduler.workers = 2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.853198] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.853398] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.853591] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.853766] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.853931] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.854111] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.854280] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.854474] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.854645] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.host_subset_size = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.854810] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.854972] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.855152] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.855321] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.isolated_hosts = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.855491] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.isolated_images = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.855654] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.855814] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.855977] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.856153] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.pci_in_placement = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.856317] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.856483] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.856645] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.856804] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.856965] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.857138] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.857303] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.track_instance_changes = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.857479] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.857651] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] metrics.required = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.857814] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] metrics.weight_multiplier = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.857977] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.858157] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] metrics.weight_setting = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.858472] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.858648] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] serial_console.enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.858823] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] serial_console.port_range = 10000:20000 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.858992] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.859177] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.859343] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] serial_console.serialproxy_port = 6083 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.859515] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_user.auth_section = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.859686] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_user.auth_type = password {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.859845] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_user.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.860038] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_user.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.860177] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_user.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.860340] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_user.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.860499] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_user.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.860668] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_user.send_service_user_token = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.860834] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_user.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.861011] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] service_user.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.861186] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.agent_enabled = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.861376] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.861698] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.861892] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.862082] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.html5proxy_port = 6082 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.862248] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.image_compression = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.862412] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.jpeg_compression = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.862573] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.playback_compression = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.862734] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.require_secure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.862901] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.server_listen = 127.0.0.1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.863081] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.863244] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.streaming_mode = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.863432] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] spice.zlib_compression = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.863606] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] upgrade_levels.baseapi = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.863777] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] upgrade_levels.compute = auto {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.863936] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] upgrade_levels.conductor = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.864110] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] upgrade_levels.scheduler = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.864284] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.864450] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.864607] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vendordata_dynamic_auth.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.864764] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vendordata_dynamic_auth.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.864925] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.865096] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vendordata_dynamic_auth.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.865257] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.865421] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.865580] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vendordata_dynamic_auth.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.865751] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.api_retry_count = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.865909] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.ca_file = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.866090] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.866262] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.cluster_name = testcl1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.866427] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.connection_pool_size = 10 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.866589] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.console_delay_seconds = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.866757] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.datastore_regex = ^datastore.* {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.866956] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.867141] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.host_password = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.867310] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.host_port = 443 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.867480] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.host_username = administrator@vsphere.local {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.867648] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.insecure = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.867808] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.integration_bridge = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.867970] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.maximum_objects = 100 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.868162] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.pbm_default_policy = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.868330] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.pbm_enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.868488] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.pbm_wsdl_location = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.868656] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.868839] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.serial_port_proxy_uri = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.869006] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.serial_port_service_uri = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.869185] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.task_poll_interval = 0.5 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.869356] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.use_linked_clone = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.869525] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.vnc_keymap = en-us {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.869689] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.vnc_port = 5900 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.869852] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vmware.vnc_port_total = 10000 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.870049] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vnc.auth_schemes = ['none'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.870231] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vnc.enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.870522] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.870706] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.870880] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vnc.novncproxy_port = 6080 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.871069] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vnc.server_listen = 127.0.0.1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.871249] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.871479] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vnc.vencrypt_ca_certs = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.871665] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vnc.vencrypt_client_cert = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.871828] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vnc.vencrypt_client_key = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.872013] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.872187] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.disable_deep_image_inspection = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.872351] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.872524] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.872687] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.872851] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.disable_rootwrap = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.873019] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.enable_numa_live_migration = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.873184] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.873382] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.873559] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.873724] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.libvirt_disable_apic = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.873885] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.874061] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.874229] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.874391] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.874554] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.874715] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.874881] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.875051] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.875216] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.875382] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.875566] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.875738] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.client_socket_timeout = 900 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.875904] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.default_pool_size = 1000 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.876082] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.keep_alive = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.876253] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.max_header_line = 16384 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.876417] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.876579] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.ssl_ca_file = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.876736] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.ssl_cert_file = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.876896] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.ssl_key_file = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.877069] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.tcp_keepidle = 600 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.877249] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.877413] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] zvm.ca_file = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.877574] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] zvm.cloud_connector_url = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.877850] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.878093] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] zvm.reachable_timeout = 300 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.878216] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_policy.enforce_new_defaults = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.878596] env[62521]: WARNING oslo_config.cfg [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 492.878782] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_policy.enforce_scope = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.878960] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_policy.policy_default_rule = default {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.879162] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.879336] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_policy.policy_file = policy.yaml {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.879515] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.879677] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.879834] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.879990] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.880164] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.880337] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.880515] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.880692] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.connection_string = messaging:// {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.880859] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.enabled = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.881037] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.es_doc_type = notification {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.881207] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.es_scroll_size = 10000 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.881405] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.es_scroll_time = 2m {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.881585] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.filter_error_trace = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.881759] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.hmac_keys = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.881928] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.sentinel_service_name = mymaster {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.882116] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.socket_timeout = 0.1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.882284] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.trace_requests = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.882449] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler.trace_sqlalchemy = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.882630] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler_jaeger.process_tags = {} {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.882791] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler_jaeger.service_name_prefix = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.882956] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] profiler_otlp.service_name_prefix = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.883135] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] remote_debug.host = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.883332] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] remote_debug.port = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.883525] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.883693] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.883862] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.884035] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.884204] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.884369] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.884530] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.884691] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.884852] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.885029] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.885192] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.885362] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.885533] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.885777] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.885904] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.886086] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.886255] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.886430] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.886596] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.886758] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.886927] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.887104] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.887272] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.887440] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.887604] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.887767] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.887930] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.888103] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.888273] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.888441] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.ssl = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.888615] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.888785] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.888948] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.889129] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.889301] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.889467] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.889656] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.889852] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_notifications.retry = -1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.890061] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.890243] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.890417] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.auth_section = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.890586] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.auth_type = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.890747] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.cafile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.890904] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.certfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.891079] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.collect_timing = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.891267] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.connect_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.891452] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.connect_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.891619] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.endpoint_id = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.891781] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.endpoint_override = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.891942] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.insecure = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.892114] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.keyfile = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.892275] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.max_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.892433] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.min_version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.892591] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.region_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.892753] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.retriable_status_codes = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.892911] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.service_name = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.893079] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.service_type = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.893245] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.split_loggers = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.893437] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.status_code_retries = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.893605] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.status_code_retry_delay = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.893789] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.timeout = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.893959] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.valid_interfaces = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.894135] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_limit.version = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.894305] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_reports.file_event_handler = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.894477] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.894637] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] oslo_reports.log_dir = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.894808] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.894967] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.895139] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.895307] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.895475] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.895632] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.895802] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.895962] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_ovs_privileged.group = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.896131] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.896299] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.896463] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.896621] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] vif_plug_ovs_privileged.user = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.896791] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.896967] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.897152] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.897324] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.897494] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.897659] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.897825] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.897989] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.898184] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.898357] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_ovs.isolate_vif = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.898528] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.898692] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.898857] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.899033] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.899201] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] os_vif_ovs.per_port_bridge = False {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.899369] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] privsep_osbrick.capabilities = [21] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.899527] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] privsep_osbrick.group = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.899684] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] privsep_osbrick.helper_command = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.899846] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.900014] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.900173] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] privsep_osbrick.user = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.900345] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.900506] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] nova_sys_admin.group = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.900663] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] nova_sys_admin.helper_command = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.900824] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.900984] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.901157] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] nova_sys_admin.user = None {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.901323] env[62521]: DEBUG oslo_service.service [None req-aabbc37a-b79f-4d45-a5c8-be5b9dbc6819 None None] ******************************************************************************** {{(pid=62521) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 492.901812] env[62521]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 493.405199] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Getting list of instances from cluster (obj){ [ 493.405199] env[62521]: value = "domain-c8" [ 493.405199] env[62521]: _type = "ClusterComputeResource" [ 493.405199] env[62521]: } {{(pid=62521) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 493.406334] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4051bb5-924a-49fd-b07a-41092a2ebfe8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 493.415442] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Got total of 0 instances {{(pid=62521) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 493.415970] env[62521]: WARNING nova.virt.vmwareapi.driver [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 493.416444] env[62521]: INFO nova.virt.node [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Generated node identity 3d21544b-bfc8-42d0-86ca-d323b5e2628f [ 493.416675] env[62521]: INFO nova.virt.node [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Wrote node identity 3d21544b-bfc8-42d0-86ca-d323b5e2628f to /opt/stack/data/n-cpu-1/compute_id [ 493.919667] env[62521]: WARNING nova.compute.manager [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Compute nodes ['3d21544b-bfc8-42d0-86ca-d323b5e2628f'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 494.926900] env[62521]: INFO nova.compute.manager [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 495.935537] env[62521]: WARNING nova.compute.manager [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 495.935537] env[62521]: DEBUG oslo_concurrency.lockutils [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.935537] env[62521]: DEBUG oslo_concurrency.lockutils [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.935537] env[62521]: DEBUG oslo_concurrency.lockutils [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 495.935537] env[62521]: DEBUG nova.compute.resource_tracker [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 495.935537] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a31517-9a7c-4f89-9875-20a03f98be07 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.944538] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a36d5f1-4e0c-4b48-b7aa-3f41a90d20f0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.959052] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0cfdde5-5b46-48bc-843b-dc5c299af0d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.965572] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4c7278-95af-4562-bf6a-2d4a6029f4b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.995193] env[62521]: DEBUG nova.compute.resource_tracker [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181433MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 495.996948] env[62521]: DEBUG oslo_concurrency.lockutils [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.996948] env[62521]: DEBUG oslo_concurrency.lockutils [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.500333] env[62521]: WARNING nova.compute.resource_tracker [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] No compute node record for cpu-1:3d21544b-bfc8-42d0-86ca-d323b5e2628f: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 3d21544b-bfc8-42d0-86ca-d323b5e2628f could not be found. [ 497.003282] env[62521]: INFO nova.compute.resource_tracker [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 3d21544b-bfc8-42d0-86ca-d323b5e2628f [ 498.511641] env[62521]: DEBUG nova.compute.resource_tracker [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 498.512186] env[62521]: DEBUG nova.compute.resource_tracker [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 498.660744] env[62521]: INFO nova.scheduler.client.report [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] [req-cc3c814e-cbbf-4974-8995-3d0e7c42befd] Created resource provider record via placement API for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 498.677420] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64208954-a7cf-4093-84f3-ce95a6ae3eed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.685249] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16201ca-dbf4-4237-a698-94f0aad614b5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.714608] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56df7792-8d52-47e4-b8c7-831ca6990eee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.721535] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee6a7e1-8969-46a5-81cb-19c7ce8d74f2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.734112] env[62521]: DEBUG nova.compute.provider_tree [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 499.269625] env[62521]: DEBUG nova.scheduler.client.report [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 499.269865] env[62521]: DEBUG nova.compute.provider_tree [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 0 to 1 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 499.270015] env[62521]: DEBUG nova.compute.provider_tree [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 499.321997] env[62521]: DEBUG nova.compute.provider_tree [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 1 to 2 during operation: update_traits {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 499.827062] env[62521]: DEBUG nova.compute.resource_tracker [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 499.827428] env[62521]: DEBUG oslo_concurrency.lockutils [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.831s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 499.827428] env[62521]: DEBUG nova.service [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Creating RPC server for service compute {{(pid=62521) start /opt/stack/nova/nova/service.py:186}} [ 499.841869] env[62521]: DEBUG nova.service [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] Join ServiceGroup membership for this service compute {{(pid=62521) start /opt/stack/nova/nova/service.py:203}} [ 499.842185] env[62521]: DEBUG nova.servicegroup.drivers.db [None req-efc64075-9188-4dbc-a8b0-f2ec42000ee8 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62521) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 533.848333] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._sync_power_states {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.358485] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Getting list of instances from cluster (obj){ [ 534.358485] env[62521]: value = "domain-c8" [ 534.358485] env[62521]: _type = "ClusterComputeResource" [ 534.358485] env[62521]: } {{(pid=62521) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 534.358485] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2285063-0cd1-4a93-ac68-70aafdbaf62c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.368046] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Got total of 0 instances {{(pid=62521) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 534.368966] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.372631] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Getting list of instances from cluster (obj){ [ 534.372631] env[62521]: value = "domain-c8" [ 534.372631] env[62521]: _type = "ClusterComputeResource" [ 534.372631] env[62521]: } {{(pid=62521) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 534.372631] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7810c4a4-1ce9-4690-96c4-678994826802 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.379204] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Got total of 0 instances {{(pid=62521) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 535.633091] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "4f3a0a33-43c0-42fb-8660-198ff4a8529a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.633370] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "4f3a0a33-43c0-42fb-8660-198ff4a8529a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.135589] env[62521]: DEBUG nova.compute.manager [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 536.686581] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.686862] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.688734] env[62521]: INFO nova.compute.claims [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.754082] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6812ee3c-b37f-4932-a10b-4eaa3ebd9a60 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.767100] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4d7c1e-07ed-47d9-aa76-e815bdb3bc38 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.808068] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22111b55-82dd-456e-9479-07906391a72a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.815708] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc23220-16c4-4f36-90a5-d663771a4f02 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.830137] env[62521]: DEBUG nova.compute.provider_tree [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.335504] env[62521]: DEBUG nova.scheduler.client.report [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 538.844391] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.157s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.845051] env[62521]: DEBUG nova.compute.manager [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 539.007297] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Acquiring lock "b51682d1-e510-4573-af69-e6dc2528b323" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.007297] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Lock "b51682d1-e510-4573-af69-e6dc2528b323" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.350324] env[62521]: DEBUG nova.compute.utils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 539.353053] env[62521]: DEBUG nova.compute.manager [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Not allocating networking since 'none' was specified. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 539.508996] env[62521]: DEBUG nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 539.856535] env[62521]: DEBUG nova.compute.manager [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 540.039127] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.039414] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.041944] env[62521]: INFO nova.compute.claims [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 540.096619] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquiring lock "3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.096871] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.599511] env[62521]: DEBUG nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 540.869919] env[62521]: DEBUG nova.compute.manager [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 541.141551] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.152209] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "bc94294f-206a-4f54-87af-2248c5ed1ba4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.152209] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "bc94294f-206a-4f54-87af-2248c5ed1ba4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.185035] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e55f5e1-8810-4b46-af25-4861c87c8d53 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.198044] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1fcac4-5244-48b7-aff6-3a14603bb0cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.206714] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.206957] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.207085] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.207603] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.207931] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.209252] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.209252] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.209252] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.210761] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.210761] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.211697] env[62521]: DEBUG nova.virt.hardware [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.212507] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86f75e2-c548-4c61-be4d-4c61030d0c21 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.253450] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2ebc80-20ce-4aad-8b1f-cbb0aad9e2b0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.258274] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997449a1-c041-4b67-8b6e-32c7df533c36 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.267375] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3a4b6f-d81b-4fe9-bcca-3ba55e09bb9d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.282234] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c028ef62-8dd4-4818-a208-e57da63ab45c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.301039] env[62521]: DEBUG nova.compute.provider_tree [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.312735] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 541.320588] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 541.320734] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-80e4fff6-f6f6-4629-b22d-545cb4779b77 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.335348] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Created folder: OpenStack in parent group-v4. [ 541.337220] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Creating folder: Project (b8d2c36e29d546e1b2130721edd8c3cb). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 541.337220] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46da129c-88b5-4e6a-8f23-3f0d7f02207f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.348768] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Created folder: Project (b8d2c36e29d546e1b2130721edd8c3cb) in parent group-v282025. [ 541.348768] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Creating folder: Instances. Parent ref: group-v282026. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 541.348768] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a614535e-1a12-48f2-b8da-160781c00de3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.359228] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Created folder: Instances in parent group-v282026. [ 541.359228] env[62521]: DEBUG oslo.service.loopingcall [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.359228] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 541.359228] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2dd02c5d-cf6d-4374-9130-48d845c737eb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.377723] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 541.377723] env[62521]: value = "task-1318327" [ 541.377723] env[62521]: _type = "Task" [ 541.377723] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.388904] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318327, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.448474] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Acquiring lock "0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.448704] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Lock "0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.656882] env[62521]: DEBUG nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.806040] env[62521]: DEBUG nova.scheduler.client.report [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.892178] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318327, 'name': CreateVM_Task, 'duration_secs': 0.369454} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.892442] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 541.895082] env[62521]: DEBUG oslo_vmware.service [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75eb49ed-f468-493f-9223-e4d234feecc2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.902366] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.903266] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.903980] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 541.904559] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73dec16f-bff3-439e-afea-26a33f92c010 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.910466] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 541.910466] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522f0262-ceec-a21b-ee6a-3d9a3ab71019" [ 541.910466] env[62521]: _type = "Task" [ 541.910466] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.924459] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522f0262-ceec-a21b-ee6a-3d9a3ab71019, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.951498] env[62521]: DEBUG nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.189711] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.262187] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Acquiring lock "71c6b4ec-60a5-4c8d-a529-a841e1b36a53" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.264291] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Lock "71c6b4ec-60a5-4c8d-a529-a841e1b36a53" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.309746] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.310335] env[62521]: DEBUG nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 542.314351] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.173s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.316111] env[62521]: INFO nova.compute.claims [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 542.422640] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.422784] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 542.423071] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.423172] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.423545] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 542.423824] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a81a901-2bc3-43ab-ba6b-ff70b895c5b4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.445188] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 542.445188] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 542.445935] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5982c941-2a56-41c0-8938-22488570344a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.453014] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fe7d88c-944e-4e26-8572-36d406870a36 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.458704] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 542.458704] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52de964d-3d96-2076-411b-674c49396c21" [ 542.458704] env[62521]: _type = "Task" [ 542.458704] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.474245] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52de964d-3d96-2076-411b-674c49396c21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.481746] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.767585] env[62521]: DEBUG nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.823898] env[62521]: DEBUG nova.compute.utils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.825181] env[62521]: DEBUG nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 542.825466] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 542.975224] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Preparing fetch location {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 542.975224] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Creating directory with path [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 542.975224] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f91a4ce-5e5c-43f6-826e-60ac6bae2389 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.007493] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Created directory with path [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 543.007493] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Fetch image to [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 543.007493] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Downloading image file data 58bd9a24-a0a4-4846-b299-475fa0f1d05d to [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk on the data store datastore1 {{(pid=62521) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 543.007493] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6b0f96-51e5-4896-8394-8f955ea9595f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.023756] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3290f4-8379-47e5-a98c-c7d1c9294fff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.038621] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23d4457-bc6c-4aed-aef3-c6eec8f84359 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.087342] env[62521]: DEBUG nova.policy [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb953bb3fb6145168e508a89fbb35bc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa1dc496c24147789a64b336d9b8bbd0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 543.089448] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978665fe-009c-4abb-8386-541e646ba380 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.097141] env[62521]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-17138ecc-e721-4980-8958-d345638264ee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.189175] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Downloading image file data 58bd9a24-a0a4-4846-b299-475fa0f1d05d to the data store datastore1 {{(pid=62521) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 543.228419] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Acquiring lock "c3a43b3a-cb59-4431-a5d2-4c96db16337f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.228660] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Lock "c3a43b3a-cb59-4431-a5d2-4c96db16337f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.275008] env[62521]: DEBUG oslo_vmware.rw_handles [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62521) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 543.340102] env[62521]: DEBUG nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 543.348968] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.506695] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a568d2-3bce-477f-b865-7d37fb646c9a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.522393] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb6b92d-21ab-47f2-9be2-93a46da2a572 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.565999] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa23ffaf-bb34-4e9d-9d43-3e9e5265950c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.578457] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9eb4e42-cb3e-48a4-a834-bdd39ea10fd4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.597032] env[62521]: DEBUG nova.compute.provider_tree [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.731762] env[62521]: DEBUG nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 543.905209] env[62521]: DEBUG oslo_vmware.rw_handles [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Completed reading data from the image iterator. {{(pid=62521) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 543.905209] env[62521]: DEBUG oslo_vmware.rw_handles [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 543.965620] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Successfully created port: a050e802-c938-4171-a45b-d4f94a2564f4 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 544.038861] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Downloaded image file data 58bd9a24-a0a4-4846-b299-475fa0f1d05d to vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk on the data store datastore1 {{(pid=62521) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 544.042403] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Caching image {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 544.042658] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Copying Virtual Disk [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk to [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 544.043686] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce80585e-444b-4151-a5f7-56101d65c171 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.058669] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 544.058669] env[62521]: value = "task-1318328" [ 544.058669] env[62521]: _type = "Task" [ 544.058669] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.068185] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318328, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.102548] env[62521]: DEBUG nova.scheduler.client.report [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 544.263549] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.357700] env[62521]: DEBUG nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 544.393050] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 544.393354] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 544.393472] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.393726] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 544.393979] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.394995] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 544.395284] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 544.395468] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 544.395677] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 544.395853] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 544.396033] env[62521]: DEBUG nova.virt.hardware [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 544.397023] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30d3c53-58cc-459b-8888-a9976ff45c1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.407905] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0afea30f-4f7f-48b1-8b27-513298761a1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.511611] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquiring lock "dba06415-d503-4420-825e-76afc59f6153" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.511853] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "dba06415-d503-4420-825e-76afc59f6153" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.570555] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318328, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.613768] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.616062] env[62521]: DEBUG nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 544.618856] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.429s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.622056] env[62521]: INFO nova.compute.claims [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 545.021379] env[62521]: DEBUG nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 545.070686] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318328, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679297} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.074199] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Copied Virtual Disk [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk to [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 545.074199] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleting the datastore file [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 545.074199] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c998e95-b1d7-43eb-9cf6-c4844b59a623 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.082271] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 545.082271] env[62521]: value = "task-1318329" [ 545.082271] env[62521]: _type = "Task" [ 545.082271] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.098611] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318329, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.126059] env[62521]: DEBUG nova.compute.utils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 545.132055] env[62521]: DEBUG nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 545.132431] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 545.231536] env[62521]: DEBUG nova.policy [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2acbdbd60b934cee9fcb1f8c79e3d644', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '01c5a629543a43cea19d0d14b5267919', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 545.546290] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.594086] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318329, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024978} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.594086] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 545.594086] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Moving file from [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009/58bd9a24-a0a4-4846-b299-475fa0f1d05d to [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d. {{(pid=62521) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 545.595532] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-5ebab4b3-31d6-414a-941f-eaea59add719 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.602688] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 545.602688] env[62521]: value = "task-1318330" [ 545.602688] env[62521]: _type = "Task" [ 545.602688] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.610574] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318330, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.640300] env[62521]: DEBUG nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 545.833956] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df7a8d1-032d-44cd-a7a6-90da13022486 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.847938] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914ec19a-4118-43f1-97ca-9ee401d0cf80 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.886716] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795b7ca4-6cbd-48aa-903e-af567f8d71f1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.894807] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4d3a5b-e08e-49ba-a3fe-aefc7836b3ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.910134] env[62521]: DEBUG nova.compute.provider_tree [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.980039] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquiring lock "df69e966-4580-42c5-8c88-53008807c1d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.980595] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "df69e966-4580-42c5-8c88-53008807c1d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.122033] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318330, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.02923} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.122033] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] File moved {{(pid=62521) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 546.122033] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Cleaning up location [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009 {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 546.122033] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleting the datastore file [datastore1] vmware_temp/c7c28b91-59e4-4666-afb8-352cc0abd009 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 546.122033] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3cf27f1b-d3be-4eef-8a55-eb988fb0321c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.127574] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 546.127574] env[62521]: value = "task-1318331" [ 546.127574] env[62521]: _type = "Task" [ 546.127574] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.139057] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318331, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.414636] env[62521]: DEBUG nova.scheduler.client.report [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.484728] env[62521]: DEBUG nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.584379] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Successfully created port: 05241c37-a746-44c2-ac0d-ac45f086c8d6 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 546.639708] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318331, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025703} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.641266] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 546.641907] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac268dcb-3f5c-4ce2-982b-2f6297986ee7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.647074] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 546.647074] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5249c08f-51c9-5044-979e-7c4a11bdf04a" [ 546.647074] env[62521]: _type = "Task" [ 546.647074] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.656095] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5249c08f-51c9-5044-979e-7c4a11bdf04a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.657137] env[62521]: DEBUG nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 546.706610] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 546.706840] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 546.706998] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 546.707192] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 546.707332] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 546.707473] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 546.707704] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 546.707863] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 546.708183] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 546.708390] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 546.708605] env[62521]: DEBUG nova.virt.hardware [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 546.709895] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ce7769-5f32-4f33-b124-4c76e7d18edb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.722395] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4632270-7194-419a-81bd-098c82a48a31 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.887268] env[62521]: ERROR nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a050e802-c938-4171-a45b-d4f94a2564f4, please check neutron logs for more information. [ 546.887268] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 546.887268] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.887268] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 546.887268] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.887268] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 546.887268] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.887268] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 546.887268] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.887268] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 546.887268] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.887268] env[62521]: ERROR nova.compute.manager raise self.value [ 546.887268] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.887268] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 546.887268] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.887268] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 546.887768] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.887768] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 546.887768] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a050e802-c938-4171-a45b-d4f94a2564f4, please check neutron logs for more information. [ 546.887768] env[62521]: ERROR nova.compute.manager [ 546.887768] env[62521]: Traceback (most recent call last): [ 546.887768] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 546.887768] env[62521]: listener.cb(fileno) [ 546.887768] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.887768] env[62521]: result = function(*args, **kwargs) [ 546.887768] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 546.887768] env[62521]: return func(*args, **kwargs) [ 546.887768] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.887768] env[62521]: raise e [ 546.887768] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.887768] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 546.887768] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.887768] env[62521]: created_port_ids = self._update_ports_for_instance( [ 546.887768] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.887768] env[62521]: with excutils.save_and_reraise_exception(): [ 546.887768] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.887768] env[62521]: self.force_reraise() [ 546.887768] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.887768] env[62521]: raise self.value [ 546.887768] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.887768] env[62521]: updated_port = self._update_port( [ 546.887768] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.887768] env[62521]: _ensure_no_port_binding_failure(port) [ 546.887768] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.887768] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 546.888643] env[62521]: nova.exception.PortBindingFailed: Binding failed for port a050e802-c938-4171-a45b-d4f94a2564f4, please check neutron logs for more information. [ 546.888643] env[62521]: Removing descriptor: 15 [ 546.889482] env[62521]: ERROR nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a050e802-c938-4171-a45b-d4f94a2564f4, please check neutron logs for more information. [ 546.889482] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] Traceback (most recent call last): [ 546.889482] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 546.889482] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] yield resources [ 546.889482] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 546.889482] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] self.driver.spawn(context, instance, image_meta, [ 546.889482] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 546.889482] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] self._vmops.spawn(context, instance, image_meta, injected_files, [ 546.889482] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 546.889482] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] vm_ref = self.build_virtual_machine(instance, [ 546.889482] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] vif_infos = vmwarevif.get_vif_info(self._session, [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] for vif in network_info: [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] return self._sync_wrapper(fn, *args, **kwargs) [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] self.wait() [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] self[:] = self._gt.wait() [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] return self._exit_event.wait() [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 546.890974] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] result = hub.switch() [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] return self.greenlet.switch() [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] result = function(*args, **kwargs) [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] return func(*args, **kwargs) [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] raise e [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] nwinfo = self.network_api.allocate_for_instance( [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] created_port_ids = self._update_ports_for_instance( [ 546.891423] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] with excutils.save_and_reraise_exception(): [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] self.force_reraise() [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] raise self.value [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] updated_port = self._update_port( [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] _ensure_no_port_binding_failure(port) [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] raise exception.PortBindingFailed(port_id=port['id']) [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] nova.exception.PortBindingFailed: Binding failed for port a050e802-c938-4171-a45b-d4f94a2564f4, please check neutron logs for more information. [ 546.891840] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] [ 546.892263] env[62521]: INFO nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Terminating instance [ 546.894979] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Acquiring lock "refresh_cache-b51682d1-e510-4573-af69-e6dc2528b323" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.894979] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Acquired lock "refresh_cache-b51682d1-e510-4573-af69-e6dc2528b323" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.894979] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.923249] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.923249] env[62521]: DEBUG nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 546.924339] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.444s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.926723] env[62521]: INFO nova.compute.claims [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.109027] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.159611] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5249c08f-51c9-5044-979e-7c4a11bdf04a, 'name': SearchDatastore_Task, 'duration_secs': 0.009306} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.159969] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.159969] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 547.160321] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a21882d-de3b-4a96-8f83-f91c95ef21c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.170059] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 547.170059] env[62521]: value = "task-1318332" [ 547.170059] env[62521]: _type = "Task" [ 547.170059] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.176987] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318332, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.437295] env[62521]: DEBUG nova.compute.utils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.440273] env[62521]: DEBUG nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 547.440273] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 547.444130] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.572910] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.575904] env[62521]: DEBUG nova.policy [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0e99fc3ead94078b185bc599ff247ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25420ed4f82e478989a011d40f037a24', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 547.600434] env[62521]: DEBUG nova.compute.manager [req-7a9b2533-0388-4596-adee-8bce379b495c req-a877ab93-084e-4c68-b0b3-bd7f84a55d20 service nova] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Received event network-changed-a050e802-c938-4171-a45b-d4f94a2564f4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.600434] env[62521]: DEBUG nova.compute.manager [req-7a9b2533-0388-4596-adee-8bce379b495c req-a877ab93-084e-4c68-b0b3-bd7f84a55d20 service nova] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Refreshing instance network info cache due to event network-changed-a050e802-c938-4171-a45b-d4f94a2564f4. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 547.600579] env[62521]: DEBUG oslo_concurrency.lockutils [req-7a9b2533-0388-4596-adee-8bce379b495c req-a877ab93-084e-4c68-b0b3-bd7f84a55d20 service nova] Acquiring lock "refresh_cache-b51682d1-e510-4573-af69-e6dc2528b323" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.684795] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318332, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.944817] env[62521]: DEBUG nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 548.080544] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Releasing lock "refresh_cache-b51682d1-e510-4573-af69-e6dc2528b323" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.080996] env[62521]: DEBUG nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 548.082602] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 548.082602] env[62521]: DEBUG oslo_concurrency.lockutils [req-7a9b2533-0388-4596-adee-8bce379b495c req-a877ab93-084e-4c68-b0b3-bd7f84a55d20 service nova] Acquired lock "refresh_cache-b51682d1-e510-4573-af69-e6dc2528b323" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.082602] env[62521]: DEBUG nova.network.neutron [req-7a9b2533-0388-4596-adee-8bce379b495c req-a877ab93-084e-4c68-b0b3-bd7f84a55d20 service nova] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Refreshing network info cache for port a050e802-c938-4171-a45b-d4f94a2564f4 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 548.082851] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe2a0c79-720b-4150-8e06-4f4d57133db6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.094345] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b310d881-55ff-41c6-87d0-391f67fad50b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.121825] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b51682d1-e510-4573-af69-e6dc2528b323 could not be found. [ 548.121825] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 548.121825] env[62521]: INFO nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Took 0.04 seconds to destroy the instance on the hypervisor. [ 548.122077] env[62521]: DEBUG oslo.service.loopingcall [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.122394] env[62521]: DEBUG nova.compute.manager [-] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.122394] env[62521]: DEBUG nova.network.neutron [-] [instance: b51682d1-e510-4573-af69-e6dc2528b323] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.176051] env[62521]: DEBUG nova.network.neutron [-] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.186332] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318332, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.724799} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.191149] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 548.191378] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 548.192224] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17aff90d-69bb-4cde-9f1a-a0bdf2facacf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.197901] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0726adaa-6f54-4ec0-a222-09c2aac8634c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.206829] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd09488-165d-4b09-8447-b908846f6536 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.214311] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 548.214311] env[62521]: value = "task-1318333" [ 548.214311] env[62521]: _type = "Task" [ 548.214311] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.247112] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b324f3-face-4651-976d-169abe6a0c1c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.256292] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318333, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.258141] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Successfully created port: 9f25c76e-cf74-4efb-9207-d77eeef09e72 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.266235] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8dcf568-e3ca-421f-a5a6-287f05fe4ea2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.283902] env[62521]: DEBUG nova.compute.provider_tree [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.327723] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.328101] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.328645] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 548.328645] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Rebuilding the list of instances to heal {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 548.627567] env[62521]: DEBUG nova.network.neutron [req-7a9b2533-0388-4596-adee-8bce379b495c req-a877ab93-084e-4c68-b0b3-bd7f84a55d20 service nova] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.679062] env[62521]: DEBUG nova.network.neutron [-] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.727984] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318333, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063624} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.727984] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 548.727984] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcce891e-9896-4f8b-b00a-6b6ff4b65630 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.753724] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 548.754085] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b8a099b-331e-4f21-951d-2c4b6c674be0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.775965] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 548.775965] env[62521]: value = "task-1318334" [ 548.775965] env[62521]: _type = "Task" [ 548.775965] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.784857] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318334, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.787826] env[62521]: DEBUG nova.scheduler.client.report [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.793435] env[62521]: DEBUG nova.network.neutron [req-7a9b2533-0388-4596-adee-8bce379b495c req-a877ab93-084e-4c68-b0b3-bd7f84a55d20 service nova] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.836015] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.836015] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.836015] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.836015] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.836015] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.836266] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Didn't find any instances for network info cache update. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 548.836332] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.836571] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.836764] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.836951] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.837288] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.837505] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.837915] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 548.837915] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.956963] env[62521]: DEBUG nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 548.994720] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.994989] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.995181] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.995376] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.995534] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.995679] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.995883] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.998415] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.998511] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.998685] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.998924] env[62521]: DEBUG nova.virt.hardware [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.000660] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2539e3ac-c196-4e91-8645-69bda1b7a3bc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.013540] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befdac33-a1e7-4dc0-b867-d90147b8ac4f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.186042] env[62521]: INFO nova.compute.manager [-] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Took 1.06 seconds to deallocate network for instance. [ 549.191057] env[62521]: DEBUG nova.compute.claims [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 549.191057] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.290866] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318334, 'name': ReconfigVM_Task, 'duration_secs': 0.294292} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.291249] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Reconfigured VM instance instance-00000001 to attach disk [datastore1] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 549.292985] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.294708] env[62521]: DEBUG nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 549.297779] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10b75673-fffc-46f5-ba94-fbad0f60acf5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.300629] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.952s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.303926] env[62521]: INFO nova.compute.claims [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 549.308795] env[62521]: DEBUG oslo_concurrency.lockutils [req-7a9b2533-0388-4596-adee-8bce379b495c req-a877ab93-084e-4c68-b0b3-bd7f84a55d20 service nova] Releasing lock "refresh_cache-b51682d1-e510-4573-af69-e6dc2528b323" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.316363] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 549.316363] env[62521]: value = "task-1318335" [ 549.316363] env[62521]: _type = "Task" [ 549.316363] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.327742] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318335, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.342659] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.423387] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquiring lock "e976d2d2-aa69-473b-9dfc-e616b0132754" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.423607] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lock "e976d2d2-aa69-473b-9dfc-e616b0132754" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.802838] env[62521]: DEBUG nova.compute.utils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 549.804109] env[62521]: DEBUG nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 549.804329] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 549.837477] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318335, 'name': Rename_Task, 'duration_secs': 0.139599} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.840463] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 549.841734] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e983b9f-1687-4975-bb4f-7e5d2d5b031a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.853254] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 549.853254] env[62521]: value = "task-1318336" [ 549.853254] env[62521]: _type = "Task" [ 549.853254] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.864713] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318336, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.927321] env[62521]: DEBUG nova.compute.manager [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.102878] env[62521]: DEBUG nova.policy [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1026d148b67b4133a8db1eb6eb024782', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aab5528ac8af49eeb8eb0e51a9d9057e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 550.182548] env[62521]: ERROR nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 05241c37-a746-44c2-ac0d-ac45f086c8d6, please check neutron logs for more information. [ 550.182548] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 550.182548] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.182548] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 550.182548] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.182548] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 550.182548] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.182548] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 550.182548] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.182548] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 550.182548] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.182548] env[62521]: ERROR nova.compute.manager raise self.value [ 550.182548] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.182548] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 550.182548] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.182548] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 550.183070] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.183070] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 550.183070] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 05241c37-a746-44c2-ac0d-ac45f086c8d6, please check neutron logs for more information. [ 550.183070] env[62521]: ERROR nova.compute.manager [ 550.183070] env[62521]: Traceback (most recent call last): [ 550.183070] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 550.183070] env[62521]: listener.cb(fileno) [ 550.183070] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.183070] env[62521]: result = function(*args, **kwargs) [ 550.183070] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 550.183070] env[62521]: return func(*args, **kwargs) [ 550.183070] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 550.183070] env[62521]: raise e [ 550.183070] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.183070] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 550.183070] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.183070] env[62521]: created_port_ids = self._update_ports_for_instance( [ 550.183070] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.183070] env[62521]: with excutils.save_and_reraise_exception(): [ 550.183070] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.183070] env[62521]: self.force_reraise() [ 550.183070] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.183070] env[62521]: raise self.value [ 550.183070] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.183070] env[62521]: updated_port = self._update_port( [ 550.183070] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.183070] env[62521]: _ensure_no_port_binding_failure(port) [ 550.183070] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.183070] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 550.183948] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 05241c37-a746-44c2-ac0d-ac45f086c8d6, please check neutron logs for more information. [ 550.183948] env[62521]: Removing descriptor: 17 [ 550.183948] env[62521]: ERROR nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 05241c37-a746-44c2-ac0d-ac45f086c8d6, please check neutron logs for more information. [ 550.183948] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Traceback (most recent call last): [ 550.183948] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 550.183948] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] yield resources [ 550.183948] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 550.183948] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] self.driver.spawn(context, instance, image_meta, [ 550.183948] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 550.183948] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.183948] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.183948] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] vm_ref = self.build_virtual_machine(instance, [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] for vif in network_info: [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] return self._sync_wrapper(fn, *args, **kwargs) [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] self.wait() [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] self[:] = self._gt.wait() [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] return self._exit_event.wait() [ 550.184298] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] result = hub.switch() [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] return self.greenlet.switch() [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] result = function(*args, **kwargs) [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] return func(*args, **kwargs) [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] raise e [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] nwinfo = self.network_api.allocate_for_instance( [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.184709] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] created_port_ids = self._update_ports_for_instance( [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] with excutils.save_and_reraise_exception(): [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] self.force_reraise() [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] raise self.value [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] updated_port = self._update_port( [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] _ensure_no_port_binding_failure(port) [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.185277] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] raise exception.PortBindingFailed(port_id=port['id']) [ 550.185672] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] nova.exception.PortBindingFailed: Binding failed for port 05241c37-a746-44c2-ac0d-ac45f086c8d6, please check neutron logs for more information. [ 550.185672] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] [ 550.185672] env[62521]: INFO nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Terminating instance [ 550.190950] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquiring lock "refresh_cache-3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.191230] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquired lock "refresh_cache-3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.191795] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 550.310092] env[62521]: DEBUG nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 550.365955] env[62521]: DEBUG oslo_vmware.api [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318336, 'name': PowerOnVM_Task, 'duration_secs': 0.46439} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.366177] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 550.366618] env[62521]: INFO nova.compute.manager [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Took 9.50 seconds to spawn the instance on the hypervisor. [ 550.366618] env[62521]: DEBUG nova.compute.manager [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 550.370956] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12796c23-6eb2-4cde-8b2b-2b890661e018 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.457742] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.617893] env[62521]: DEBUG nova.compute.manager [req-2476d1a5-b590-423c-86f7-81c791679f05 req-46ccf6fb-5f0e-470a-bd71-4d9c50e5aa52 service nova] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Received event network-vif-deleted-a050e802-c938-4171-a45b-d4f94a2564f4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.618591] env[62521]: DEBUG nova.compute.manager [req-2476d1a5-b590-423c-86f7-81c791679f05 req-46ccf6fb-5f0e-470a-bd71-4d9c50e5aa52 service nova] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Received event network-changed-05241c37-a746-44c2-ac0d-ac45f086c8d6 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.618591] env[62521]: DEBUG nova.compute.manager [req-2476d1a5-b590-423c-86f7-81c791679f05 req-46ccf6fb-5f0e-470a-bd71-4d9c50e5aa52 service nova] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Refreshing instance network info cache due to event network-changed-05241c37-a746-44c2-ac0d-ac45f086c8d6. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 550.618591] env[62521]: DEBUG oslo_concurrency.lockutils [req-2476d1a5-b590-423c-86f7-81c791679f05 req-46ccf6fb-5f0e-470a-bd71-4d9c50e5aa52 service nova] Acquiring lock "refresh_cache-3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.679249] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bc0c3a-5a5a-4a1d-b94d-eb0a0f862107 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.690911] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a3b6ad5-eb2c-450d-b788-601d5a9fa882 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.726704] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f038184-2cc8-4544-a2eb-f9a046d5d434 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.736636] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb72d9b-9300-4c15-949e-0476e7896885 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.750811] env[62521]: DEBUG nova.compute.provider_tree [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.769651] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.892129] env[62521]: INFO nova.compute.manager [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Took 14.25 seconds to build instance. [ 551.046204] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Acquiring lock "da34c648-4074-494e-a569-e8e839207a70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.046448] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Lock "da34c648-4074-494e-a569-e8e839207a70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.098734] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.257020] env[62521]: DEBUG nova.scheduler.client.report [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.322293] env[62521]: DEBUG nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 551.355215] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 551.355483] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 551.355717] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.355809] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 551.355940] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.356447] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 551.357026] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 551.357444] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 551.357728] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 551.358876] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 551.359963] env[62521]: DEBUG nova.virt.hardware [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 551.360392] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a92cbb6-71bb-445f-874b-e5bec151055b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.370483] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a2cb8f-56d4-4495-b282-0d5ad3c66ed1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.376702] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Successfully created port: bfd3f27e-25d5-445e-8451-d2511d8c40f3 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.394674] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29478113-fae8-4ab9-a451-04487524dc3d tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "4f3a0a33-43c0-42fb-8660-198ff4a8529a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.761s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.602072] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Releasing lock "refresh_cache-3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.602512] env[62521]: DEBUG nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 551.602709] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 551.603029] env[62521]: DEBUG oslo_concurrency.lockutils [req-2476d1a5-b590-423c-86f7-81c791679f05 req-46ccf6fb-5f0e-470a-bd71-4d9c50e5aa52 service nova] Acquired lock "refresh_cache-3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.603291] env[62521]: DEBUG nova.network.neutron [req-2476d1a5-b590-423c-86f7-81c791679f05 req-46ccf6fb-5f0e-470a-bd71-4d9c50e5aa52 service nova] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Refreshing network info cache for port 05241c37-a746-44c2-ac0d-ac45f086c8d6 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 551.604411] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-558c10b3-122c-4671-afb5-a34dbaef2d43 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.612924] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc8b9fb-94a4-49dc-a587-334c20ab3143 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.635953] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a could not be found. [ 551.636096] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 551.638420] env[62521]: INFO nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 551.638420] env[62521]: DEBUG oslo.service.loopingcall [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.638420] env[62521]: DEBUG nova.compute.manager [-] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 551.638420] env[62521]: DEBUG nova.network.neutron [-] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 551.695448] env[62521]: DEBUG nova.network.neutron [-] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.760776] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.760963] env[62521]: DEBUG nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 551.764152] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.501s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.765637] env[62521]: INFO nova.compute.claims [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.898095] env[62521]: DEBUG nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 551.909190] env[62521]: ERROR nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9f25c76e-cf74-4efb-9207-d77eeef09e72, please check neutron logs for more information. [ 551.909190] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.909190] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.909190] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.909190] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.909190] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.909190] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.909190] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.909190] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.909190] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 551.909190] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.909190] env[62521]: ERROR nova.compute.manager raise self.value [ 551.909190] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.909190] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.909190] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.909190] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.910866] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.910866] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.910866] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9f25c76e-cf74-4efb-9207-d77eeef09e72, please check neutron logs for more information. [ 551.910866] env[62521]: ERROR nova.compute.manager [ 551.910866] env[62521]: Traceback (most recent call last): [ 551.910866] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.910866] env[62521]: listener.cb(fileno) [ 551.910866] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.910866] env[62521]: result = function(*args, **kwargs) [ 551.910866] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.910866] env[62521]: return func(*args, **kwargs) [ 551.910866] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.910866] env[62521]: raise e [ 551.910866] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.910866] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 551.910866] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.910866] env[62521]: created_port_ids = self._update_ports_for_instance( [ 551.910866] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.910866] env[62521]: with excutils.save_and_reraise_exception(): [ 551.910866] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.910866] env[62521]: self.force_reraise() [ 551.910866] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.910866] env[62521]: raise self.value [ 551.910866] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.910866] env[62521]: updated_port = self._update_port( [ 551.910866] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.910866] env[62521]: _ensure_no_port_binding_failure(port) [ 551.910866] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.910866] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.912100] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 9f25c76e-cf74-4efb-9207-d77eeef09e72, please check neutron logs for more information. [ 551.912100] env[62521]: Removing descriptor: 18 [ 551.912100] env[62521]: ERROR nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9f25c76e-cf74-4efb-9207-d77eeef09e72, please check neutron logs for more information. [ 551.912100] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Traceback (most recent call last): [ 551.912100] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 551.912100] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] yield resources [ 551.912100] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.912100] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] self.driver.spawn(context, instance, image_meta, [ 551.912100] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 551.912100] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.912100] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.912100] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] vm_ref = self.build_virtual_machine(instance, [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] for vif in network_info: [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] return self._sync_wrapper(fn, *args, **kwargs) [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] self.wait() [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] self[:] = self._gt.wait() [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] return self._exit_event.wait() [ 551.912432] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] result = hub.switch() [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] return self.greenlet.switch() [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] result = function(*args, **kwargs) [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] return func(*args, **kwargs) [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] raise e [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] nwinfo = self.network_api.allocate_for_instance( [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.913203] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] created_port_ids = self._update_ports_for_instance( [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] with excutils.save_and_reraise_exception(): [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] self.force_reraise() [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] raise self.value [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] updated_port = self._update_port( [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] _ensure_no_port_binding_failure(port) [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.913609] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] raise exception.PortBindingFailed(port_id=port['id']) [ 551.914043] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] nova.exception.PortBindingFailed: Binding failed for port 9f25c76e-cf74-4efb-9207-d77eeef09e72, please check neutron logs for more information. [ 551.914043] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] [ 551.914043] env[62521]: INFO nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Terminating instance [ 551.914043] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "refresh_cache-bc94294f-206a-4f54-87af-2248c5ed1ba4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.914043] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "refresh_cache-bc94294f-206a-4f54-87af-2248c5ed1ba4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.914043] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.166570] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "20b01df0-4ced-4668-a786-046251a0c7ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.167060] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "20b01df0-4ced-4668-a786-046251a0c7ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.180472] env[62521]: DEBUG nova.network.neutron [req-2476d1a5-b590-423c-86f7-81c791679f05 req-46ccf6fb-5f0e-470a-bd71-4d9c50e5aa52 service nova] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.195679] env[62521]: DEBUG nova.network.neutron [-] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.270251] env[62521]: DEBUG nova.compute.utils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 552.273894] env[62521]: DEBUG nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 552.273894] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 552.307525] env[62521]: DEBUG nova.network.neutron [req-2476d1a5-b590-423c-86f7-81c791679f05 req-46ccf6fb-5f0e-470a-bd71-4d9c50e5aa52 service nova] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.404811] env[62521]: DEBUG nova.policy [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '83b26d6fcd6e4c6f954c157f018e270c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b7c5c95d37e74c30946425ee28a22ff4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 552.438186] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.447559] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.597420] env[62521]: DEBUG nova.compute.manager [req-01bb1351-2de2-4957-ae48-1b4193c8c6e7 req-5eb97341-945e-4dfd-b670-e085e2cef00e service nova] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Received event network-changed-9f25c76e-cf74-4efb-9207-d77eeef09e72 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 552.597618] env[62521]: DEBUG nova.compute.manager [req-01bb1351-2de2-4957-ae48-1b4193c8c6e7 req-5eb97341-945e-4dfd-b670-e085e2cef00e service nova] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Refreshing instance network info cache due to event network-changed-9f25c76e-cf74-4efb-9207-d77eeef09e72. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 552.597841] env[62521]: DEBUG oslo_concurrency.lockutils [req-01bb1351-2de2-4957-ae48-1b4193c8c6e7 req-5eb97341-945e-4dfd-b670-e085e2cef00e service nova] Acquiring lock "refresh_cache-bc94294f-206a-4f54-87af-2248c5ed1ba4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.690808] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.702240] env[62521]: INFO nova.compute.manager [-] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Took 1.06 seconds to deallocate network for instance. [ 552.707510] env[62521]: DEBUG nova.compute.claims [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 552.708891] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.779352] env[62521]: DEBUG nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 552.815092] env[62521]: DEBUG oslo_concurrency.lockutils [req-2476d1a5-b590-423c-86f7-81c791679f05 req-46ccf6fb-5f0e-470a-bd71-4d9c50e5aa52 service nova] Releasing lock "refresh_cache-3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.028726] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3a69b8-5c04-4541-a04b-66027aae94fe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.037299] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b60ab6d-5784-4c91-b1c3-d06b8f015dbc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.074630] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f963625-c09f-4b10-98c3-1e1629a4687e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.085235] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c643b349-7dca-4ef2-a01b-ad32cb247e20 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.097056] env[62521]: DEBUG nova.compute.provider_tree [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.194769] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "refresh_cache-bc94294f-206a-4f54-87af-2248c5ed1ba4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.195719] env[62521]: DEBUG nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 553.195719] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 553.195996] env[62521]: DEBUG oslo_concurrency.lockutils [req-01bb1351-2de2-4957-ae48-1b4193c8c6e7 req-5eb97341-945e-4dfd-b670-e085e2cef00e service nova] Acquired lock "refresh_cache-bc94294f-206a-4f54-87af-2248c5ed1ba4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.196881] env[62521]: DEBUG nova.network.neutron [req-01bb1351-2de2-4957-ae48-1b4193c8c6e7 req-5eb97341-945e-4dfd-b670-e085e2cef00e service nova] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Refreshing network info cache for port 9f25c76e-cf74-4efb-9207-d77eeef09e72 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 553.197739] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0bf556a0-5ad6-4118-9b08-42bb36eb3918 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.213622] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f19755-cdc7-4845-9826-cd7b49838a1d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.240017] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bc94294f-206a-4f54-87af-2248c5ed1ba4 could not be found. [ 553.240400] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 553.240710] env[62521]: INFO nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 553.241079] env[62521]: DEBUG oslo.service.loopingcall [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 553.241570] env[62521]: DEBUG nova.compute.manager [-] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 553.241805] env[62521]: DEBUG nova.network.neutron [-] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 553.467047] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Successfully created port: f943804f-fca5-4c9b-96da-fe5bd2df773d {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 553.475533] env[62521]: DEBUG nova.network.neutron [-] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.601810] env[62521]: DEBUG nova.scheduler.client.report [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.620138] env[62521]: DEBUG nova.compute.manager [req-e9672b07-02da-4519-b219-1e538b1a1071 req-87f6d89b-62fc-47a0-9476-8dc57528f480 service nova] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Received event network-vif-deleted-05241c37-a746-44c2-ac0d-ac45f086c8d6 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 553.736177] env[62521]: DEBUG nova.network.neutron [req-01bb1351-2de2-4957-ae48-1b4193c8c6e7 req-5eb97341-945e-4dfd-b670-e085e2cef00e service nova] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.800432] env[62521]: DEBUG nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 553.832747] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.833041] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.833407] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.833407] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.833532] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.833635] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.833906] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.834316] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.834576] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.835085] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.835085] env[62521]: DEBUG nova.virt.hardware [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.836032] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b50d94a-7584-40c2-9d1a-b2c6520aa6a7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.844601] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd1467c-9f24-4d18-bada-83ed44ed053b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.877203] env[62521]: DEBUG nova.network.neutron [req-01bb1351-2de2-4957-ae48-1b4193c8c6e7 req-5eb97341-945e-4dfd-b670-e085e2cef00e service nova] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.979125] env[62521]: DEBUG nova.network.neutron [-] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.009071] env[62521]: INFO nova.compute.manager [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Rebuilding instance [ 554.069117] env[62521]: DEBUG nova.compute.manager [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 554.070136] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e429564a-974b-4213-a4e6-91c84454d647 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.107986] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.107986] env[62521]: DEBUG nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 554.111469] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.568s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.114827] env[62521]: INFO nova.compute.claims [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 554.379417] env[62521]: DEBUG oslo_concurrency.lockutils [req-01bb1351-2de2-4957-ae48-1b4193c8c6e7 req-5eb97341-945e-4dfd-b670-e085e2cef00e service nova] Releasing lock "refresh_cache-bc94294f-206a-4f54-87af-2248c5ed1ba4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.483269] env[62521]: INFO nova.compute.manager [-] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Took 1.24 seconds to deallocate network for instance. [ 554.486263] env[62521]: DEBUG nova.compute.claims [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 554.489148] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.592258] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 554.592658] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4aa08db2-445d-4d4e-87b2-038d463e91cf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.601594] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 554.601594] env[62521]: value = "task-1318337" [ 554.601594] env[62521]: _type = "Task" [ 554.601594] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.610975] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318337, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.620695] env[62521]: DEBUG nova.compute.utils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 554.630123] env[62521]: DEBUG nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 554.630123] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 554.767338] env[62521]: DEBUG nova.policy [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0310c36a6d5e4738b821dbca13aa1049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4790f943d139446a85bb0fcb412a4341', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 555.119627] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318337, 'name': PowerOffVM_Task, 'duration_secs': 0.157186} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.120750] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 555.120750] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 555.121510] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e53b10-9af8-46f1-ad15-fb7e28aacdb7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.129611] env[62521]: DEBUG nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 555.132517] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 555.138804] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22cd1c76-3c99-49e5-b83b-ef68284d480d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.206100] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 555.206359] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 555.207528] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleting the datastore file [datastore1] 4f3a0a33-43c0-42fb-8660-198ff4a8529a {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 555.207528] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-533ced93-1b60-4aaa-aba1-e86fcef83890 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.218753] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 555.218753] env[62521]: value = "task-1318339" [ 555.218753] env[62521]: _type = "Task" [ 555.218753] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.230395] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318339, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.368840] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a81f8a4-0e38-40cd-a1c0-c7e4040c753e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.376595] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2434112-927c-4449-8106-fddc46396223 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.419053] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35cde419-cc52-4c6b-930e-86f52e6b3c5e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.428837] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce3106b3-3563-4d3a-9a56-4c7209cc3411 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.444929] env[62521]: DEBUG nova.compute.provider_tree [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.737274] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318339, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143647} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.738406] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 555.738904] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 555.741071] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 555.901892] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Successfully created port: 7e1b61ae-32ea-4647-9046-23f7241687c2 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 555.904641] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Acquiring lock "6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.904849] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Lock "6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.949036] env[62521]: DEBUG nova.scheduler.client.report [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.033664] env[62521]: ERROR nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bfd3f27e-25d5-445e-8451-d2511d8c40f3, please check neutron logs for more information. [ 556.033664] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.033664] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.033664] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.033664] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.033664] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.033664] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.033664] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.033664] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.033664] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 556.033664] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.033664] env[62521]: ERROR nova.compute.manager raise self.value [ 556.033664] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.033664] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.033664] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.033664] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.034281] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.034281] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.034281] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bfd3f27e-25d5-445e-8451-d2511d8c40f3, please check neutron logs for more information. [ 556.034281] env[62521]: ERROR nova.compute.manager [ 556.034281] env[62521]: Traceback (most recent call last): [ 556.034281] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.034281] env[62521]: listener.cb(fileno) [ 556.034281] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.034281] env[62521]: result = function(*args, **kwargs) [ 556.034281] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 556.034281] env[62521]: return func(*args, **kwargs) [ 556.034281] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.034281] env[62521]: raise e [ 556.034281] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.034281] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 556.034281] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.034281] env[62521]: created_port_ids = self._update_ports_for_instance( [ 556.034281] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.034281] env[62521]: with excutils.save_and_reraise_exception(): [ 556.034281] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.034281] env[62521]: self.force_reraise() [ 556.034281] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.034281] env[62521]: raise self.value [ 556.034281] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.034281] env[62521]: updated_port = self._update_port( [ 556.034281] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.034281] env[62521]: _ensure_no_port_binding_failure(port) [ 556.034281] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.034281] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.035199] env[62521]: nova.exception.PortBindingFailed: Binding failed for port bfd3f27e-25d5-445e-8451-d2511d8c40f3, please check neutron logs for more information. [ 556.035199] env[62521]: Removing descriptor: 15 [ 556.035199] env[62521]: ERROR nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bfd3f27e-25d5-445e-8451-d2511d8c40f3, please check neutron logs for more information. [ 556.035199] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Traceback (most recent call last): [ 556.035199] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 556.035199] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] yield resources [ 556.035199] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.035199] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] self.driver.spawn(context, instance, image_meta, [ 556.035199] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 556.035199] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.035199] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.035199] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] vm_ref = self.build_virtual_machine(instance, [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] for vif in network_info: [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] return self._sync_wrapper(fn, *args, **kwargs) [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] self.wait() [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] self[:] = self._gt.wait() [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] return self._exit_event.wait() [ 556.035559] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] result = hub.switch() [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] return self.greenlet.switch() [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] result = function(*args, **kwargs) [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] return func(*args, **kwargs) [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] raise e [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] nwinfo = self.network_api.allocate_for_instance( [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.035906] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] created_port_ids = self._update_ports_for_instance( [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] with excutils.save_and_reraise_exception(): [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] self.force_reraise() [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] raise self.value [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] updated_port = self._update_port( [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] _ensure_no_port_binding_failure(port) [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.036289] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] raise exception.PortBindingFailed(port_id=port['id']) [ 556.036637] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] nova.exception.PortBindingFailed: Binding failed for port bfd3f27e-25d5-445e-8451-d2511d8c40f3, please check neutron logs for more information. [ 556.036637] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] [ 556.036637] env[62521]: INFO nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Terminating instance [ 556.038009] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Acquiring lock "refresh_cache-0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.038154] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Acquired lock "refresh_cache-0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.038634] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 556.139914] env[62521]: DEBUG nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 556.178911] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.179156] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.179310] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.179484] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.179622] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.179759] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.179954] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.183221] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.183547] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.183757] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.183958] env[62521]: DEBUG nova.virt.hardware [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.185258] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff757dc-585f-4ece-bf1b-9c9788b6ab40 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.198451] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f19083f-66a6-4b98-983a-5dc0a0178dd3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.227157] env[62521]: DEBUG nova.compute.manager [req-c6801c55-03cd-4f11-9b70-bba2260336b3 req-2bda8de8-84c4-4fb8-bc17-79a55938576b service nova] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Received event network-vif-deleted-9f25c76e-cf74-4efb-9207-d77eeef09e72 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 556.454631] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.455261] env[62521]: DEBUG nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 556.459138] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.354s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.460488] env[62521]: INFO nova.compute.claims [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 556.471059] env[62521]: DEBUG nova.compute.manager [req-e3e0435b-8e15-4b9d-8a00-61ae2e790047 req-90e01e7f-9ad5-4727-8f8e-97d133cd3536 service nova] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Received event network-changed-bfd3f27e-25d5-445e-8451-d2511d8c40f3 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 556.471247] env[62521]: DEBUG nova.compute.manager [req-e3e0435b-8e15-4b9d-8a00-61ae2e790047 req-90e01e7f-9ad5-4727-8f8e-97d133cd3536 service nova] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Refreshing instance network info cache due to event network-changed-bfd3f27e-25d5-445e-8451-d2511d8c40f3. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 556.471432] env[62521]: DEBUG oslo_concurrency.lockutils [req-e3e0435b-8e15-4b9d-8a00-61ae2e790047 req-90e01e7f-9ad5-4727-8f8e-97d133cd3536 service nova] Acquiring lock "refresh_cache-0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.587447] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.748321] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.795587] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.795870] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.795972] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.796181] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.796336] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.796484] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.796692] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.796849] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.797214] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.797442] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.797760] env[62521]: DEBUG nova.virt.hardware [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.800266] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3414f5-9ff3-4182-abb4-04eb92e20505 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.811416] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f154731-2b0c-4e86-9b19-94cb003a7a0d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.827100] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 556.834030] env[62521]: DEBUG oslo.service.loopingcall [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 556.834969] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 556.835108] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5a4a529-cf59-482c-a2bd-78f52b32b7a2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.854724] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 556.854724] env[62521]: value = "task-1318340" [ 556.854724] env[62521]: _type = "Task" [ 556.854724] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.869100] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318340, 'name': CreateVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.960953] env[62521]: DEBUG nova.compute.utils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 556.963668] env[62521]: DEBUG nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 556.964100] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 557.027584] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquiring lock "7653f469-5ce5-4d74-9a3f-55a6adb3a255" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.027796] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lock "7653f469-5ce5-4d74-9a3f-55a6adb3a255" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.132763] env[62521]: DEBUG nova.policy [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e265e90b68a44f2ea47fd3828733758e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28071af632ef48bd87a2ccfe842c206d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 557.256895] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Releasing lock "refresh_cache-0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.256895] env[62521]: DEBUG nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 557.256895] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 557.256895] env[62521]: DEBUG oslo_concurrency.lockutils [req-e3e0435b-8e15-4b9d-8a00-61ae2e790047 req-90e01e7f-9ad5-4727-8f8e-97d133cd3536 service nova] Acquired lock "refresh_cache-0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.256895] env[62521]: DEBUG nova.network.neutron [req-e3e0435b-8e15-4b9d-8a00-61ae2e790047 req-90e01e7f-9ad5-4727-8f8e-97d133cd3536 service nova] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Refreshing network info cache for port bfd3f27e-25d5-445e-8451-d2511d8c40f3 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 557.257508] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-deb4bb4b-09f5-4411-a2b6-abc4a9224e97 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.274147] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a196027-b48f-40e6-933a-49c50c8ddbcc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.303258] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745 could not be found. [ 557.304683] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 557.304683] env[62521]: INFO nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Took 0.05 seconds to destroy the instance on the hypervisor. [ 557.304683] env[62521]: DEBUG oslo.service.loopingcall [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.304683] env[62521]: DEBUG nova.compute.manager [-] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.304683] env[62521]: DEBUG nova.network.neutron [-] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 557.369213] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318340, 'name': CreateVM_Task, 'duration_secs': 0.285189} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.369760] env[62521]: DEBUG nova.network.neutron [-] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.374333] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 557.375776] env[62521]: DEBUG oslo_vmware.service [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f64e1d3-337b-48c0-bc11-d243b400d675 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.381897] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.382070] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.382704] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 557.382959] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f887ac01-c75f-4fe7-8f06-fc9a381c4a93 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.394743] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 557.394743] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5273cde9-f6f5-e258-0daf-e2a91af4c852" [ 557.394743] env[62521]: _type = "Task" [ 557.394743] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.406773] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5273cde9-f6f5-e258-0daf-e2a91af4c852, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.471671] env[62521]: DEBUG nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 557.746867] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef24942-36fb-430c-b938-2a487b15f777 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.760594] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5cb10ad-8f66-45e0-a3c8-516e5dab7635 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.803911] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a34dee-859f-473d-83c2-b103ae27c399 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.811446] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4d4c49-91a3-4644-9956-b442666ca3e0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.827936] env[62521]: DEBUG nova.compute.provider_tree [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.846942] env[62521]: DEBUG nova.network.neutron [req-e3e0435b-8e15-4b9d-8a00-61ae2e790047 req-90e01e7f-9ad5-4727-8f8e-97d133cd3536 service nova] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.861305] env[62521]: ERROR nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f943804f-fca5-4c9b-96da-fe5bd2df773d, please check neutron logs for more information. [ 557.861305] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 557.861305] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.861305] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 557.861305] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.861305] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 557.861305] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.861305] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 557.861305] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.861305] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 557.861305] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.861305] env[62521]: ERROR nova.compute.manager raise self.value [ 557.861305] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.861305] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 557.861305] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.861305] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 557.861799] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.861799] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 557.861799] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f943804f-fca5-4c9b-96da-fe5bd2df773d, please check neutron logs for more information. [ 557.861799] env[62521]: ERROR nova.compute.manager [ 557.861799] env[62521]: Traceback (most recent call last): [ 557.861799] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 557.861799] env[62521]: listener.cb(fileno) [ 557.861799] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.861799] env[62521]: result = function(*args, **kwargs) [ 557.861799] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.861799] env[62521]: return func(*args, **kwargs) [ 557.861799] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.861799] env[62521]: raise e [ 557.861799] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.861799] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 557.861799] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.861799] env[62521]: created_port_ids = self._update_ports_for_instance( [ 557.861799] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.861799] env[62521]: with excutils.save_and_reraise_exception(): [ 557.861799] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.861799] env[62521]: self.force_reraise() [ 557.861799] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.861799] env[62521]: raise self.value [ 557.861799] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.861799] env[62521]: updated_port = self._update_port( [ 557.861799] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.861799] env[62521]: _ensure_no_port_binding_failure(port) [ 557.861799] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.861799] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 557.865671] env[62521]: nova.exception.PortBindingFailed: Binding failed for port f943804f-fca5-4c9b-96da-fe5bd2df773d, please check neutron logs for more information. [ 557.865671] env[62521]: Removing descriptor: 17 [ 557.865671] env[62521]: ERROR nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f943804f-fca5-4c9b-96da-fe5bd2df773d, please check neutron logs for more information. [ 557.865671] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Traceback (most recent call last): [ 557.865671] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 557.865671] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] yield resources [ 557.865671] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.865671] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] self.driver.spawn(context, instance, image_meta, [ 557.865671] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 557.865671] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.865671] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.865671] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] vm_ref = self.build_virtual_machine(instance, [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] for vif in network_info: [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] return self._sync_wrapper(fn, *args, **kwargs) [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] self.wait() [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] self[:] = self._gt.wait() [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] return self._exit_event.wait() [ 557.866691] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] result = hub.switch() [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] return self.greenlet.switch() [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] result = function(*args, **kwargs) [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] return func(*args, **kwargs) [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] raise e [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] nwinfo = self.network_api.allocate_for_instance( [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.867099] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] created_port_ids = self._update_ports_for_instance( [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] with excutils.save_and_reraise_exception(): [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] self.force_reraise() [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] raise self.value [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] updated_port = self._update_port( [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] _ensure_no_port_binding_failure(port) [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.867421] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] raise exception.PortBindingFailed(port_id=port['id']) [ 557.867720] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] nova.exception.PortBindingFailed: Binding failed for port f943804f-fca5-4c9b-96da-fe5bd2df773d, please check neutron logs for more information. [ 557.867720] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] [ 557.867720] env[62521]: INFO nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Terminating instance [ 557.867720] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Acquiring lock "refresh_cache-71c6b4ec-60a5-4c8d-a529-a841e1b36a53" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.867720] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Acquired lock "refresh_cache-71c6b4ec-60a5-4c8d-a529-a841e1b36a53" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.867720] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.879624] env[62521]: DEBUG nova.network.neutron [-] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.909654] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.909654] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 557.909930] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.910010] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.910225] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 557.910520] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ec3ba24-47f1-44b5-9d9f-b0b030e8274b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.919217] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 557.920048] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 557.920367] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c529e5e-aabc-448d-ad6e-92e8813f56fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.928609] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dcf9b70-a733-4d40-b6e0-70a080b1cee6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.936470] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 557.936470] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5290b25e-eec6-e006-4a0f-81b525d5b51f" [ 557.936470] env[62521]: _type = "Task" [ 557.936470] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.947080] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5290b25e-eec6-e006-4a0f-81b525d5b51f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.043678] env[62521]: DEBUG nova.network.neutron [req-e3e0435b-8e15-4b9d-8a00-61ae2e790047 req-90e01e7f-9ad5-4727-8f8e-97d133cd3536 service nova] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.225242] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Successfully created port: d12d3198-135b-4dc6-aafe-d02cca92d79e {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 558.331915] env[62521]: DEBUG nova.scheduler.client.report [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 558.386420] env[62521]: INFO nova.compute.manager [-] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Took 1.08 seconds to deallocate network for instance. [ 558.389754] env[62521]: DEBUG nova.compute.claims [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 558.389754] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.433101] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.452098] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Preparing fetch location {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 558.452633] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Creating directory with path [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 558.454199] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-526b96eb-7e5e-4f0b-9d0b-46de53cc09ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.489677] env[62521]: DEBUG nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 558.503655] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Created directory with path [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 558.503655] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Fetch image to [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 558.503655] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Downloading image file data 58bd9a24-a0a4-4846-b299-475fa0f1d05d to [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk on the data store datastore2 {{(pid=62521) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 558.503655] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f77aed-d5e1-40d3-a0d9-75ab33ec0156 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.509728] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b261bda-cd8f-40ac-88fb-621f40201a47 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.517312] env[62521]: ERROR nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7e1b61ae-32ea-4647-9046-23f7241687c2, please check neutron logs for more information. [ 558.517312] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 558.517312] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.517312] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 558.517312] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.517312] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 558.517312] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.517312] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 558.517312] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.517312] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 558.517312] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.517312] env[62521]: ERROR nova.compute.manager raise self.value [ 558.517312] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.517312] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 558.517312] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.517312] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 558.517863] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.517863] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 558.517863] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7e1b61ae-32ea-4647-9046-23f7241687c2, please check neutron logs for more information. [ 558.517863] env[62521]: ERROR nova.compute.manager [ 558.517863] env[62521]: Traceback (most recent call last): [ 558.517863] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 558.517863] env[62521]: listener.cb(fileno) [ 558.517863] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.517863] env[62521]: result = function(*args, **kwargs) [ 558.517863] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 558.517863] env[62521]: return func(*args, **kwargs) [ 558.517863] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.517863] env[62521]: raise e [ 558.517863] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.517863] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 558.517863] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.517863] env[62521]: created_port_ids = self._update_ports_for_instance( [ 558.517863] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.517863] env[62521]: with excutils.save_and_reraise_exception(): [ 558.517863] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.517863] env[62521]: self.force_reraise() [ 558.517863] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.517863] env[62521]: raise self.value [ 558.517863] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.517863] env[62521]: updated_port = self._update_port( [ 558.517863] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.517863] env[62521]: _ensure_no_port_binding_failure(port) [ 558.517863] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.517863] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 558.519277] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 7e1b61ae-32ea-4647-9046-23f7241687c2, please check neutron logs for more information. [ 558.519277] env[62521]: Removing descriptor: 18 [ 558.528056] env[62521]: ERROR nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7e1b61ae-32ea-4647-9046-23f7241687c2, please check neutron logs for more information. [ 558.528056] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Traceback (most recent call last): [ 558.528056] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 558.528056] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] yield resources [ 558.528056] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 558.528056] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] self.driver.spawn(context, instance, image_meta, [ 558.528056] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 558.528056] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.528056] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.528056] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] vm_ref = self.build_virtual_machine(instance, [ 558.528056] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] for vif in network_info: [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] return self._sync_wrapper(fn, *args, **kwargs) [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] self.wait() [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] self[:] = self._gt.wait() [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] return self._exit_event.wait() [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.528461] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] result = hub.switch() [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] return self.greenlet.switch() [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] result = function(*args, **kwargs) [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] return func(*args, **kwargs) [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] raise e [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] nwinfo = self.network_api.allocate_for_instance( [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] created_port_ids = self._update_ports_for_instance( [ 558.528812] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] with excutils.save_and_reraise_exception(): [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] self.force_reraise() [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] raise self.value [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] updated_port = self._update_port( [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] _ensure_no_port_binding_failure(port) [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] raise exception.PortBindingFailed(port_id=port['id']) [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] nova.exception.PortBindingFailed: Binding failed for port 7e1b61ae-32ea-4647-9046-23f7241687c2, please check neutron logs for more information. [ 558.529250] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] [ 558.529640] env[62521]: INFO nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Terminating instance [ 558.530911] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e46a4d6-5491-4b9c-b626-12d513ca2c69 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.535811] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Acquiring lock "refresh_cache-c3a43b3a-cb59-4431-a5d2-4c96db16337f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.535979] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Acquired lock "refresh_cache-c3a43b3a-cb59-4431-a5d2-4c96db16337f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.536159] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.539150] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 558.539980] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 558.539980] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 558.539980] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 558.539980] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 558.539980] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 558.540193] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 558.540385] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 558.540625] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 558.540777] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 558.540951] env[62521]: DEBUG nova.virt.hardware [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 558.542161] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29aeddb2-8de2-4b03-a19f-56daf35b9423 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.571313] env[62521]: DEBUG oslo_concurrency.lockutils [req-e3e0435b-8e15-4b9d-8a00-61ae2e790047 req-90e01e7f-9ad5-4727-8f8e-97d133cd3536 service nova] Releasing lock "refresh_cache-0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.575453] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b39f42c9-df78-41aa-a788-855913b2bd33 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.583421] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2254b96-5197-4de2-adb4-211e409a4723 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.590539] env[62521]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-223935bf-b002-4296-9573-b058d7f91869 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.625229] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Downloading image file data 58bd9a24-a0a4-4846-b299-475fa0f1d05d to the data store datastore2 {{(pid=62521) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 558.703045] env[62521]: DEBUG oslo_vmware.rw_handles [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62521) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 558.839879] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.840666] env[62521]: DEBUG nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 558.848426] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.658s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.910586] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.951778] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquiring lock "6296d4ea-d678-4981-bc5a-729e83020fa2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.952196] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "6296d4ea-d678-4981-bc5a-729e83020fa2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.162900] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.353208] env[62521]: DEBUG nova.compute.utils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 559.357885] env[62521]: DEBUG nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 559.358045] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 559.403890] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.417880] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Releasing lock "refresh_cache-71c6b4ec-60a5-4c8d-a529-a841e1b36a53" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.419379] env[62521]: DEBUG nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.420335] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.420667] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-675392cf-0527-4263-b29e-c5aa5cd643c1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.434994] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5110a389-0368-4772-b1e9-af4f876dd70a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.452975] env[62521]: DEBUG oslo_vmware.rw_handles [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Completed reading data from the image iterator. {{(pid=62521) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 559.453170] env[62521]: DEBUG oslo_vmware.rw_handles [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 559.467157] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 71c6b4ec-60a5-4c8d-a529-a841e1b36a53 could not be found. [ 559.467396] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.467881] env[62521]: INFO nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Took 0.05 seconds to destroy the instance on the hypervisor. [ 559.468212] env[62521]: DEBUG oslo.service.loopingcall [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.473868] env[62521]: DEBUG nova.policy [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ceb79989e28f4cd0a2cb3671837965f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '531419d03627425da3d3713038b7e54e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 559.473937] env[62521]: DEBUG nova.compute.manager [-] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.474661] env[62521]: DEBUG nova.network.neutron [-] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.489486] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Downloaded image file data 58bd9a24-a0a4-4846-b299-475fa0f1d05d to vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk on the data store datastore2 {{(pid=62521) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 559.491934] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Caching image {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 559.492130] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Copying Virtual Disk [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk to [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 559.492647] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c670600-32ed-4f38-8939-00c13b07dc4b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.508019] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 559.508019] env[62521]: value = "task-1318345" [ 559.508019] env[62521]: _type = "Task" [ 559.508019] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.515811] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318345, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.520257] env[62521]: DEBUG nova.network.neutron [-] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.635949] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "4b2fda93-a60d-4050-ae05-7a37ac284917" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.637980] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "4b2fda93-a60d-4050-ae05-7a37ac284917" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.675524] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "1d25cdca-3c87-47a2-85aa-86f7132297b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.675912] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "1d25cdca-3c87-47a2-85aa-86f7132297b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.704081] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14669aae-3554-4d3d-9eeb-7186d641cfec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.714504] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d71550d-14ca-44c0-b326-5062a7cb7053 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.752550] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ddc2cb-0505-43e8-815b-877fe59aa180 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.760544] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789d464b-99eb-4b11-afb7-1f48227ebea3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.775961] env[62521]: DEBUG nova.compute.provider_tree [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.860051] env[62521]: DEBUG nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 559.913822] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Releasing lock "refresh_cache-c3a43b3a-cb59-4431-a5d2-4c96db16337f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.914156] env[62521]: DEBUG nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.914361] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.917878] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0425a977-a72e-4ad8-a027-da5fbebb3e85 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.928728] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e53c007c-9358-4e82-9ef5-3ecd279e6384 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.954115] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c3a43b3a-cb59-4431-a5d2-4c96db16337f could not be found. [ 559.954397] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.954525] env[62521]: INFO nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 559.954779] env[62521]: DEBUG oslo.service.loopingcall [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.954980] env[62521]: DEBUG nova.compute.manager [-] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.955088] env[62521]: DEBUG nova.network.neutron [-] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 560.001653] env[62521]: DEBUG nova.network.neutron [-] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.021552] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318345, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.024052] env[62521]: DEBUG nova.network.neutron [-] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.101894] env[62521]: DEBUG nova.compute.manager [req-5806ecb8-412b-434b-9943-fc0ddd488fc2 req-3276a7de-47fa-4c3e-b9e9-28a565c752fa service nova] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Received event network-changed-7e1b61ae-32ea-4647-9046-23f7241687c2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.101894] env[62521]: DEBUG nova.compute.manager [req-5806ecb8-412b-434b-9943-fc0ddd488fc2 req-3276a7de-47fa-4c3e-b9e9-28a565c752fa service nova] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Refreshing instance network info cache due to event network-changed-7e1b61ae-32ea-4647-9046-23f7241687c2. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 560.102730] env[62521]: DEBUG oslo_concurrency.lockutils [req-5806ecb8-412b-434b-9943-fc0ddd488fc2 req-3276a7de-47fa-4c3e-b9e9-28a565c752fa service nova] Acquiring lock "refresh_cache-c3a43b3a-cb59-4431-a5d2-4c96db16337f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.102730] env[62521]: DEBUG oslo_concurrency.lockutils [req-5806ecb8-412b-434b-9943-fc0ddd488fc2 req-3276a7de-47fa-4c3e-b9e9-28a565c752fa service nova] Acquired lock "refresh_cache-c3a43b3a-cb59-4431-a5d2-4c96db16337f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.102730] env[62521]: DEBUG nova.network.neutron [req-5806ecb8-412b-434b-9943-fc0ddd488fc2 req-3276a7de-47fa-4c3e-b9e9-28a565c752fa service nova] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Refreshing network info cache for port 7e1b61ae-32ea-4647-9046-23f7241687c2 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 560.265940] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Successfully created port: 2fb2566d-d73e-4378-b78e-29be75da5f59 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 560.278937] env[62521]: DEBUG nova.scheduler.client.report [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.504037] env[62521]: DEBUG nova.network.neutron [-] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.518154] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318345, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669674} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.518413] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Copied Virtual Disk [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk to [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 560.518590] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleting the datastore file [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d/tmp-sparse.vmdk {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 560.518826] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5319d332-346a-4ea8-8730-332a56db5e91 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.525175] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 560.525175] env[62521]: value = "task-1318346" [ 560.525175] env[62521]: _type = "Task" [ 560.525175] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.528646] env[62521]: INFO nova.compute.manager [-] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Took 1.05 seconds to deallocate network for instance. [ 560.531444] env[62521]: DEBUG nova.compute.claims [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.531617] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.535827] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.674527] env[62521]: DEBUG nova.network.neutron [req-5806ecb8-412b-434b-9943-fc0ddd488fc2 req-3276a7de-47fa-4c3e-b9e9-28a565c752fa service nova] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.786832] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.938s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.788632] env[62521]: ERROR nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a050e802-c938-4171-a45b-d4f94a2564f4, please check neutron logs for more information. [ 560.788632] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] Traceback (most recent call last): [ 560.788632] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 560.788632] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] self.driver.spawn(context, instance, image_meta, [ 560.788632] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 560.788632] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.788632] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.788632] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] vm_ref = self.build_virtual_machine(instance, [ 560.788632] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.788632] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.788632] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] for vif in network_info: [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] return self._sync_wrapper(fn, *args, **kwargs) [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] self.wait() [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] self[:] = self._gt.wait() [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] return self._exit_event.wait() [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] result = hub.switch() [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.789124] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] return self.greenlet.switch() [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] result = function(*args, **kwargs) [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] return func(*args, **kwargs) [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] raise e [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] nwinfo = self.network_api.allocate_for_instance( [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] created_port_ids = self._update_ports_for_instance( [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] with excutils.save_and_reraise_exception(): [ 560.789517] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] self.force_reraise() [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] raise self.value [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] updated_port = self._update_port( [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] _ensure_no_port_binding_failure(port) [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] raise exception.PortBindingFailed(port_id=port['id']) [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] nova.exception.PortBindingFailed: Binding failed for port a050e802-c938-4171-a45b-d4f94a2564f4, please check neutron logs for more information. [ 560.789972] env[62521]: ERROR nova.compute.manager [instance: b51682d1-e510-4573-af69-e6dc2528b323] [ 560.790289] env[62521]: DEBUG nova.compute.utils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Binding failed for port a050e802-c938-4171-a45b-d4f94a2564f4, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 560.793667] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 11.450s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.793911] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.794180] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 560.794893] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.340s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.798108] env[62521]: INFO nova.compute.claims [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 560.807575] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e70c2f-d5fe-459e-919f-4f71df3e2de1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.822940] env[62521]: DEBUG nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Build of instance b51682d1-e510-4573-af69-e6dc2528b323 was re-scheduled: Binding failed for port a050e802-c938-4171-a45b-d4f94a2564f4, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 560.824281] env[62521]: DEBUG nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 560.824725] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Acquiring lock "refresh_cache-b51682d1-e510-4573-af69-e6dc2528b323" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.827633] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Acquired lock "refresh_cache-b51682d1-e510-4573-af69-e6dc2528b323" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.827633] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.839521] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9c0813-ce21-4054-ba74-53b47fa1e6a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.874821] env[62521]: DEBUG nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 560.882968] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991e8f39-3b93-44d0-9b38-5c0eca386d8b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.897433] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3681d04-99c8-42b7-83d8-c7dc4d896ec8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.901562] env[62521]: DEBUG nova.network.neutron [req-5806ecb8-412b-434b-9943-fc0ddd488fc2 req-3276a7de-47fa-4c3e-b9e9-28a565c752fa service nova] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.934155] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181440MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 560.934322] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.939210] env[62521]: DEBUG nova.compute.manager [req-96781b6f-5213-492b-8e25-359fcbf72de1 req-6d84c5e5-2c25-4437-9741-c570f37e0675 service nova] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Received event network-vif-deleted-bfd3f27e-25d5-445e-8451-d2511d8c40f3 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.939345] env[62521]: DEBUG nova.compute.manager [req-96781b6f-5213-492b-8e25-359fcbf72de1 req-6d84c5e5-2c25-4437-9741-c570f37e0675 service nova] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Received event network-changed-f943804f-fca5-4c9b-96da-fe5bd2df773d {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.939499] env[62521]: DEBUG nova.compute.manager [req-96781b6f-5213-492b-8e25-359fcbf72de1 req-6d84c5e5-2c25-4437-9741-c570f37e0675 service nova] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Refreshing instance network info cache due to event network-changed-f943804f-fca5-4c9b-96da-fe5bd2df773d. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 560.939702] env[62521]: DEBUG oslo_concurrency.lockutils [req-96781b6f-5213-492b-8e25-359fcbf72de1 req-6d84c5e5-2c25-4437-9741-c570f37e0675 service nova] Acquiring lock "refresh_cache-71c6b4ec-60a5-4c8d-a529-a841e1b36a53" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.940300] env[62521]: DEBUG oslo_concurrency.lockutils [req-96781b6f-5213-492b-8e25-359fcbf72de1 req-6d84c5e5-2c25-4437-9741-c570f37e0675 service nova] Acquired lock "refresh_cache-71c6b4ec-60a5-4c8d-a529-a841e1b36a53" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.940300] env[62521]: DEBUG nova.network.neutron [req-96781b6f-5213-492b-8e25-359fcbf72de1 req-6d84c5e5-2c25-4437-9741-c570f37e0675 service nova] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Refreshing network info cache for port f943804f-fca5-4c9b-96da-fe5bd2df773d {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 560.950112] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 560.950352] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 560.950504] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 560.950678] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 560.950820] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 560.950975] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 560.951178] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 560.951330] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 560.951492] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 560.951646] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 560.952028] env[62521]: DEBUG nova.virt.hardware [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 560.953389] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc1bc17-457b-47b6-8b9b-e0cea25f887c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.962602] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c876b39c-1c55-4c7b-95ff-3e3b3431e3cb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.007088] env[62521]: INFO nova.compute.manager [-] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Took 1.05 seconds to deallocate network for instance. [ 561.013007] env[62521]: DEBUG nova.compute.claims [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 561.013142] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.035114] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022222} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.035367] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 561.035567] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Moving file from [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b/58bd9a24-a0a4-4846-b299-475fa0f1d05d to [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d. {{(pid=62521) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 561.035812] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-ddd47062-0e0d-4300-9e26-231700d11101 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.042294] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 561.042294] env[62521]: value = "task-1318348" [ 561.042294] env[62521]: _type = "Task" [ 561.042294] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.051912] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318348, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.379410] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.381600] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquiring lock "1c892e78-a863-4561-940d-b44701e97041" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.381862] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "1c892e78-a863-4561-940d-b44701e97041" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.404310] env[62521]: DEBUG oslo_concurrency.lockutils [req-5806ecb8-412b-434b-9943-fc0ddd488fc2 req-3276a7de-47fa-4c3e-b9e9-28a565c752fa service nova] Releasing lock "refresh_cache-c3a43b3a-cb59-4431-a5d2-4c96db16337f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.496894] env[62521]: DEBUG nova.network.neutron [req-96781b6f-5213-492b-8e25-359fcbf72de1 req-6d84c5e5-2c25-4437-9741-c570f37e0675 service nova] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.559344] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318348, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024295} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.559344] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] File moved {{(pid=62521) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 561.559344] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Cleaning up location [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 561.559344] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleting the datastore file [datastore2] vmware_temp/b1e14dfe-ed8d-48d2-973f-7f96f669681b {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 561.559344] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c064868-741c-4351-885e-630d2a6bf45a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.568031] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 561.568031] env[62521]: value = "task-1318349" [ 561.568031] env[62521]: _type = "Task" [ 561.568031] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.579557] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318349, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.581611] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.750780] env[62521]: DEBUG nova.network.neutron [req-96781b6f-5213-492b-8e25-359fcbf72de1 req-6d84c5e5-2c25-4437-9741-c570f37e0675 service nova] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.017737] env[62521]: ERROR nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d12d3198-135b-4dc6-aafe-d02cca92d79e, please check neutron logs for more information. [ 562.017737] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 562.017737] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.017737] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 562.017737] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.017737] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 562.017737] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.017737] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 562.017737] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.017737] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 562.017737] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.017737] env[62521]: ERROR nova.compute.manager raise self.value [ 562.017737] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.017737] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 562.017737] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.017737] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 562.018498] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.018498] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 562.018498] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d12d3198-135b-4dc6-aafe-d02cca92d79e, please check neutron logs for more information. [ 562.018498] env[62521]: ERROR nova.compute.manager [ 562.018498] env[62521]: Traceback (most recent call last): [ 562.018498] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 562.018498] env[62521]: listener.cb(fileno) [ 562.018498] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.018498] env[62521]: result = function(*args, **kwargs) [ 562.018498] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 562.018498] env[62521]: return func(*args, **kwargs) [ 562.018498] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.018498] env[62521]: raise e [ 562.018498] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.018498] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 562.018498] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.018498] env[62521]: created_port_ids = self._update_ports_for_instance( [ 562.018498] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.018498] env[62521]: with excutils.save_and_reraise_exception(): [ 562.018498] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.018498] env[62521]: self.force_reraise() [ 562.018498] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.018498] env[62521]: raise self.value [ 562.018498] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.018498] env[62521]: updated_port = self._update_port( [ 562.018498] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.018498] env[62521]: _ensure_no_port_binding_failure(port) [ 562.018498] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.018498] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 562.019325] env[62521]: nova.exception.PortBindingFailed: Binding failed for port d12d3198-135b-4dc6-aafe-d02cca92d79e, please check neutron logs for more information. [ 562.019325] env[62521]: Removing descriptor: 15 [ 562.019325] env[62521]: ERROR nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d12d3198-135b-4dc6-aafe-d02cca92d79e, please check neutron logs for more information. [ 562.019325] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] Traceback (most recent call last): [ 562.019325] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 562.019325] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] yield resources [ 562.019325] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 562.019325] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] self.driver.spawn(context, instance, image_meta, [ 562.019325] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 562.019325] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.019325] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.019325] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] vm_ref = self.build_virtual_machine(instance, [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] for vif in network_info: [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] return self._sync_wrapper(fn, *args, **kwargs) [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] self.wait() [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] self[:] = self._gt.wait() [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] return self._exit_event.wait() [ 562.019722] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] result = hub.switch() [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] return self.greenlet.switch() [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] result = function(*args, **kwargs) [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] return func(*args, **kwargs) [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] raise e [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] nwinfo = self.network_api.allocate_for_instance( [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.020144] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] created_port_ids = self._update_ports_for_instance( [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] with excutils.save_and_reraise_exception(): [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] self.force_reraise() [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] raise self.value [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] updated_port = self._update_port( [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] _ensure_no_port_binding_failure(port) [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.020494] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] raise exception.PortBindingFailed(port_id=port['id']) [ 562.020809] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] nova.exception.PortBindingFailed: Binding failed for port d12d3198-135b-4dc6-aafe-d02cca92d79e, please check neutron logs for more information. [ 562.020809] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] [ 562.020809] env[62521]: INFO nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Terminating instance [ 562.026147] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquiring lock "refresh_cache-dba06415-d503-4420-825e-76afc59f6153" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.026147] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquired lock "refresh_cache-dba06415-d503-4420-825e-76afc59f6153" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.026313] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.084601] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Releasing lock "refresh_cache-b51682d1-e510-4573-af69-e6dc2528b323" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.084891] env[62521]: DEBUG nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 562.085080] env[62521]: DEBUG nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.085248] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 562.087099] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318349, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033073} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.087288] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 562.088323] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ac60092-5bf7-43b2-8515-8acfe4d40cca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.097155] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 562.097155] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ac95d3-a6c8-a374-a8f8-d5d3660587b9" [ 562.097155] env[62521]: _type = "Task" [ 562.097155] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.106216] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ac95d3-a6c8-a374-a8f8-d5d3660587b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.150283] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.186612] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bd932c-1ba0-41ea-87c1-4b253cace032 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.194823] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec75eab-c185-4e4b-9baf-4241fa34ca50 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.242844] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05fa16d0-e521-4a18-abdb-a44931100c1b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.252354] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03fc0bd-6362-476e-b59a-f334802d5353 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.256260] env[62521]: DEBUG oslo_concurrency.lockutils [req-96781b6f-5213-492b-8e25-359fcbf72de1 req-6d84c5e5-2c25-4437-9741-c570f37e0675 service nova] Releasing lock "refresh_cache-71c6b4ec-60a5-4c8d-a529-a841e1b36a53" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.256260] env[62521]: DEBUG nova.compute.manager [req-96781b6f-5213-492b-8e25-359fcbf72de1 req-6d84c5e5-2c25-4437-9741-c570f37e0675 service nova] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Received event network-vif-deleted-f943804f-fca5-4c9b-96da-fe5bd2df773d {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 562.267557] env[62521]: DEBUG nova.compute.provider_tree [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.555917] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.609736] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ac95d3-a6c8-a374-a8f8-d5d3660587b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009712} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.609960] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.610242] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 562.610965] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0769ba52-10e8-4a95-a1df-f0abc44f826f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.617031] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 562.617031] env[62521]: value = "task-1318350" [ 562.617031] env[62521]: _type = "Task" [ 562.617031] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.627312] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318350, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.655465] env[62521]: DEBUG nova.network.neutron [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.774329] env[62521]: DEBUG nova.scheduler.client.report [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.779072] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.876439] env[62521]: DEBUG nova.compute.manager [req-906f7bc7-b5e2-4c8b-9372-7ce3e485eb74 req-6d3cec90-fba0-4835-aab7-88ecdb8a37f0 service nova] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Received event network-vif-deleted-7e1b61ae-32ea-4647-9046-23f7241687c2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.127997] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318350, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.161488] env[62521]: INFO nova.compute.manager [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] [instance: b51682d1-e510-4573-af69-e6dc2528b323] Took 1.08 seconds to deallocate network for instance. [ 563.281585] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.282366] env[62521]: DEBUG nova.compute.manager [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 563.285386] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.848s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.287993] env[62521]: INFO nova.compute.claims [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 563.291437] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Releasing lock "refresh_cache-dba06415-d503-4420-825e-76afc59f6153" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.291437] env[62521]: DEBUG nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 563.291437] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 563.291986] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab44006b-374b-4398-b387-aaae73a0b940 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.302982] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16001316-76c3-4a7f-be6b-fd541dd347ef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.329698] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dba06415-d503-4420-825e-76afc59f6153 could not be found. [ 563.329698] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 563.329698] env[62521]: INFO nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Took 0.04 seconds to destroy the instance on the hypervisor. [ 563.329698] env[62521]: DEBUG oslo.service.loopingcall [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.329698] env[62521]: DEBUG nova.compute.manager [-] [instance: dba06415-d503-4420-825e-76afc59f6153] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.329698] env[62521]: DEBUG nova.network.neutron [-] [instance: dba06415-d503-4420-825e-76afc59f6153] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.346245] env[62521]: DEBUG nova.network.neutron [-] [instance: dba06415-d503-4420-825e-76afc59f6153] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.596493] env[62521]: DEBUG nova.compute.manager [req-475100f9-c647-4aff-9785-036db3039da1 req-115da0bf-f12b-44c9-98d3-4ce7e737207e service nova] [instance: dba06415-d503-4420-825e-76afc59f6153] Received event network-changed-d12d3198-135b-4dc6-aafe-d02cca92d79e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.596816] env[62521]: DEBUG nova.compute.manager [req-475100f9-c647-4aff-9785-036db3039da1 req-115da0bf-f12b-44c9-98d3-4ce7e737207e service nova] [instance: dba06415-d503-4420-825e-76afc59f6153] Refreshing instance network info cache due to event network-changed-d12d3198-135b-4dc6-aafe-d02cca92d79e. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 563.596944] env[62521]: DEBUG oslo_concurrency.lockutils [req-475100f9-c647-4aff-9785-036db3039da1 req-115da0bf-f12b-44c9-98d3-4ce7e737207e service nova] Acquiring lock "refresh_cache-dba06415-d503-4420-825e-76afc59f6153" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.597047] env[62521]: DEBUG oslo_concurrency.lockutils [req-475100f9-c647-4aff-9785-036db3039da1 req-115da0bf-f12b-44c9-98d3-4ce7e737207e service nova] Acquired lock "refresh_cache-dba06415-d503-4420-825e-76afc59f6153" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.597223] env[62521]: DEBUG nova.network.neutron [req-475100f9-c647-4aff-9785-036db3039da1 req-115da0bf-f12b-44c9-98d3-4ce7e737207e service nova] [instance: dba06415-d503-4420-825e-76afc59f6153] Refreshing network info cache for port d12d3198-135b-4dc6-aafe-d02cca92d79e {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 563.628972] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318350, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527675} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.629249] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 563.629556] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 563.629687] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-265eb8a6-5dcb-4616-a50d-0a1abda7c2e3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.639245] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 563.639245] env[62521]: value = "task-1318352" [ 563.639245] env[62521]: _type = "Task" [ 563.639245] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.647214] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318352, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.715961] env[62521]: ERROR nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2fb2566d-d73e-4378-b78e-29be75da5f59, please check neutron logs for more information. [ 563.715961] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.715961] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.715961] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.715961] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.715961] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.715961] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.715961] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.715961] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.715961] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 563.715961] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.715961] env[62521]: ERROR nova.compute.manager raise self.value [ 563.715961] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.715961] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.715961] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.715961] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.716438] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.716438] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.716438] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2fb2566d-d73e-4378-b78e-29be75da5f59, please check neutron logs for more information. [ 563.716438] env[62521]: ERROR nova.compute.manager [ 563.716438] env[62521]: Traceback (most recent call last): [ 563.716438] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.716438] env[62521]: listener.cb(fileno) [ 563.716438] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.716438] env[62521]: result = function(*args, **kwargs) [ 563.716438] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.716438] env[62521]: return func(*args, **kwargs) [ 563.716438] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.716438] env[62521]: raise e [ 563.716438] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.716438] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 563.716438] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.716438] env[62521]: created_port_ids = self._update_ports_for_instance( [ 563.716438] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.716438] env[62521]: with excutils.save_and_reraise_exception(): [ 563.716438] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.716438] env[62521]: self.force_reraise() [ 563.716438] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.716438] env[62521]: raise self.value [ 563.716438] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.716438] env[62521]: updated_port = self._update_port( [ 563.716438] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.716438] env[62521]: _ensure_no_port_binding_failure(port) [ 563.716438] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.716438] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.717349] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 2fb2566d-d73e-4378-b78e-29be75da5f59, please check neutron logs for more information. [ 563.717349] env[62521]: Removing descriptor: 18 [ 563.717349] env[62521]: ERROR nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2fb2566d-d73e-4378-b78e-29be75da5f59, please check neutron logs for more information. [ 563.717349] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] Traceback (most recent call last): [ 563.717349] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 563.717349] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] yield resources [ 563.717349] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.717349] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] self.driver.spawn(context, instance, image_meta, [ 563.717349] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 563.717349] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.717349] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.717349] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] vm_ref = self.build_virtual_machine(instance, [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] for vif in network_info: [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] return self._sync_wrapper(fn, *args, **kwargs) [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] self.wait() [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] self[:] = self._gt.wait() [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] return self._exit_event.wait() [ 563.717834] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] result = hub.switch() [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] return self.greenlet.switch() [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] result = function(*args, **kwargs) [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] return func(*args, **kwargs) [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] raise e [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] nwinfo = self.network_api.allocate_for_instance( [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.719769] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] created_port_ids = self._update_ports_for_instance( [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] with excutils.save_and_reraise_exception(): [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] self.force_reraise() [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] raise self.value [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] updated_port = self._update_port( [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] _ensure_no_port_binding_failure(port) [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.720191] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] raise exception.PortBindingFailed(port_id=port['id']) [ 563.720536] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] nova.exception.PortBindingFailed: Binding failed for port 2fb2566d-d73e-4378-b78e-29be75da5f59, please check neutron logs for more information. [ 563.720536] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] [ 563.720536] env[62521]: INFO nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Terminating instance [ 563.720536] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquiring lock "refresh_cache-df69e966-4580-42c5-8c88-53008807c1d0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.720536] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquired lock "refresh_cache-df69e966-4580-42c5-8c88-53008807c1d0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.720536] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.797415] env[62521]: DEBUG nova.compute.utils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.799876] env[62521]: DEBUG nova.compute.manager [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Not allocating networking since 'none' was specified. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 563.849125] env[62521]: DEBUG nova.network.neutron [-] [instance: dba06415-d503-4420-825e-76afc59f6153] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.125430] env[62521]: DEBUG nova.network.neutron [req-475100f9-c647-4aff-9785-036db3039da1 req-115da0bf-f12b-44c9-98d3-4ce7e737207e service nova] [instance: dba06415-d503-4420-825e-76afc59f6153] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.153239] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318352, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068335} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.153484] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 564.154317] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ba32a9-0079-4561-8fac-421a7668451d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.178123] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 564.178123] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-554ce6f7-4b36-477f-94cc-e39729c56262 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.204059] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 564.204059] env[62521]: value = "task-1318353" [ 564.204059] env[62521]: _type = "Task" [ 564.204059] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.214871] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318353, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.224912] env[62521]: INFO nova.scheduler.client.report [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Deleted allocations for instance b51682d1-e510-4573-af69-e6dc2528b323 [ 564.258088] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.267544] env[62521]: DEBUG nova.network.neutron [req-475100f9-c647-4aff-9785-036db3039da1 req-115da0bf-f12b-44c9-98d3-4ce7e737207e service nova] [instance: dba06415-d503-4420-825e-76afc59f6153] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.304598] env[62521]: DEBUG nova.compute.manager [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 564.351978] env[62521]: INFO nova.compute.manager [-] [instance: dba06415-d503-4420-825e-76afc59f6153] Took 1.02 seconds to deallocate network for instance. [ 564.354509] env[62521]: DEBUG nova.compute.claims [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 564.354723] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.399159] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.580822] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d1eece-6630-4e56-bdaf-6299e6b285f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.591092] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb80b05e-2d9e-47b2-86c9-afc7193c7169 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.632063] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7729b3-57b0-46b8-bc3c-ba871df4c8ee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.640101] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c26123-5946-44c3-b31e-3a34aab250b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.655414] env[62521]: DEBUG nova.compute.provider_tree [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.719204] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318353, 'name': ReconfigVM_Task, 'duration_secs': 0.299287} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.719852] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Reconfigured VM instance instance-00000001 to attach disk [datastore2] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 564.720552] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-43826b9f-70a5-438b-90da-cdcbd6f8ebe8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.730922] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 564.730922] env[62521]: value = "task-1318354" [ 564.730922] env[62521]: _type = "Task" [ 564.730922] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.743215] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d7d6b93-3d55-49c3-80b7-4913f1fcaf20 tempest-ServerDiagnosticsTest-393412027 tempest-ServerDiagnosticsTest-393412027-project-member] Lock "b51682d1-e510-4573-af69-e6dc2528b323" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.737s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.743215] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318354, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.770524] env[62521]: DEBUG oslo_concurrency.lockutils [req-475100f9-c647-4aff-9785-036db3039da1 req-115da0bf-f12b-44c9-98d3-4ce7e737207e service nova] Releasing lock "refresh_cache-dba06415-d503-4420-825e-76afc59f6153" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.770640] env[62521]: DEBUG nova.compute.manager [req-475100f9-c647-4aff-9785-036db3039da1 req-115da0bf-f12b-44c9-98d3-4ce7e737207e service nova] [instance: dba06415-d503-4420-825e-76afc59f6153] Received event network-vif-deleted-d12d3198-135b-4dc6-aafe-d02cca92d79e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 564.911238] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Releasing lock "refresh_cache-df69e966-4580-42c5-8c88-53008807c1d0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.911238] env[62521]: DEBUG nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 564.911238] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.911238] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-934f1c7d-d7aa-4693-befc-8ff6253c3dd1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.919975] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4471621-2a43-418c-a880-9b52e27e14c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.949252] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance df69e966-4580-42c5-8c88-53008807c1d0 could not be found. [ 564.949380] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.949569] env[62521]: INFO nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 564.949816] env[62521]: DEBUG oslo.service.loopingcall [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.950044] env[62521]: DEBUG nova.compute.manager [-] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.950135] env[62521]: DEBUG nova.network.neutron [-] [instance: df69e966-4580-42c5-8c88-53008807c1d0] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.995134] env[62521]: DEBUG nova.network.neutron [-] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.158190] env[62521]: DEBUG nova.scheduler.client.report [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 565.240543] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318354, 'name': Rename_Task, 'duration_secs': 0.134865} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.241133] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 565.241203] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-afccf795-43b3-40fa-9a8b-27225ae9e3ef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.247020] env[62521]: DEBUG nova.compute.manager [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.248806] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 565.248806] env[62521]: value = "task-1318356" [ 565.248806] env[62521]: _type = "Task" [ 565.248806] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.258341] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318356, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.314987] env[62521]: DEBUG nova.compute.manager [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 565.344018] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.344018] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.344018] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.344280] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.344459] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.344798] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.345157] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.345448] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.345721] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.345986] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.346313] env[62521]: DEBUG nova.virt.hardware [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.348124] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839c854d-697c-46ba-bbf7-6247c2002122 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.361020] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e469a8-cd7f-4df4-8ece-d132bc311298 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.375517] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 565.383390] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Creating folder: Project (5f188aadf4524949aed68326682ec65c). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 565.383728] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3d6ae52-2df3-4408-ae62-e7ba681d05f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.394907] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Created folder: Project (5f188aadf4524949aed68326682ec65c) in parent group-v282025. [ 565.394907] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Creating folder: Instances. Parent ref: group-v282033. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 565.394907] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-017e07ed-7cf3-411e-9e5a-39cd5b299757 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.403065] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Created folder: Instances in parent group-v282033. [ 565.403432] env[62521]: DEBUG oslo.service.loopingcall [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.403626] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 565.403626] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d97e90cb-c4d9-404a-ba9e-48b8fd3c6f24 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.420127] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 565.420127] env[62521]: value = "task-1318359" [ 565.420127] env[62521]: _type = "Task" [ 565.420127] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.427621] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318359, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.502970] env[62521]: DEBUG nova.network.neutron [-] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.663941] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.664649] env[62521]: DEBUG nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 565.673027] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.961s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.772316] env[62521]: DEBUG oslo_vmware.api [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318356, 'name': PowerOnVM_Task, 'duration_secs': 0.46315} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.772662] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 565.773221] env[62521]: DEBUG nova.compute.manager [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 565.774495] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6645628-024f-481a-ba01-3d4e1acc0cd0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.784598] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.931750] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318359, 'name': CreateVM_Task, 'duration_secs': 0.30285} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.933073] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 565.933073] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.933073] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.933250] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 565.933399] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f9830ed-4112-4e37-b9e7-dfd35d1e78a3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.939058] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 565.939058] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523f8ef1-50aa-20a2-c3bb-22f98d2e61f2" [ 565.939058] env[62521]: _type = "Task" [ 565.939058] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.947467] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523f8ef1-50aa-20a2-c3bb-22f98d2e61f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.005834] env[62521]: INFO nova.compute.manager [-] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Took 1.06 seconds to deallocate network for instance. [ 566.011745] env[62521]: DEBUG nova.compute.claims [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 566.012136] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.175436] env[62521]: DEBUG nova.compute.utils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 566.181490] env[62521]: DEBUG nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 566.181672] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 566.299524] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.344543] env[62521]: DEBUG nova.policy [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dacb6b91d23a43ab959491505fade137', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd8558ede7d84e2a80833b953d4d23b5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 566.461292] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Acquiring lock "9c2c0e66-5c67-4e0a-9205-4080bda44a66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.461532] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Lock "9c2c0e66-5c67-4e0a-9205-4080bda44a66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.467333] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523f8ef1-50aa-20a2-c3bb-22f98d2e61f2, 'name': SearchDatastore_Task, 'duration_secs': 0.029402} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.467848] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.468268] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 566.468521] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.468660] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.469338] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 566.469610] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d7a6c87-f447-4763-8e08-20365c08da3a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.479118] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 566.479118] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 566.484859] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e51fc611-f444-4010-871a-c6c5ece1cdaf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.494315] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 566.494315] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52da92fa-4f6a-9404-ee61-69fa39c9dea5" [ 566.494315] env[62521]: _type = "Task" [ 566.494315] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.506448] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52da92fa-4f6a-9404-ee61-69fa39c9dea5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.565181] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Acquiring lock "da9d2279-79ae-4371-bef5-08ae15f2d9c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.566649] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Lock "da9d2279-79ae-4371-bef5-08ae15f2d9c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.570257] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71f28a1-18e5-4ffb-b7cb-16d531a2d0f0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.580089] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b58f8d0-0e01-4001-8eda-4cbc61b765c1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.616550] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66736eb4-b647-4276-a1f2-1fdc366465c4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.626415] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57d42ae-fb4d-4953-ae3b-51355dd1e666 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.640089] env[62521]: DEBUG nova.compute.provider_tree [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.682474] env[62521]: DEBUG nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 566.691663] env[62521]: DEBUG nova.compute.manager [req-23f13881-5a5b-4e02-b58f-a23740e58f76 req-bbc996de-5d18-42e9-ae0c-f9d290ad36d3 service nova] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Received event network-changed-2fb2566d-d73e-4378-b78e-29be75da5f59 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 566.691663] env[62521]: DEBUG nova.compute.manager [req-23f13881-5a5b-4e02-b58f-a23740e58f76 req-bbc996de-5d18-42e9-ae0c-f9d290ad36d3 service nova] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Refreshing instance network info cache due to event network-changed-2fb2566d-d73e-4378-b78e-29be75da5f59. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 566.691850] env[62521]: DEBUG oslo_concurrency.lockutils [req-23f13881-5a5b-4e02-b58f-a23740e58f76 req-bbc996de-5d18-42e9-ae0c-f9d290ad36d3 service nova] Acquiring lock "refresh_cache-df69e966-4580-42c5-8c88-53008807c1d0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.692782] env[62521]: DEBUG oslo_concurrency.lockutils [req-23f13881-5a5b-4e02-b58f-a23740e58f76 req-bbc996de-5d18-42e9-ae0c-f9d290ad36d3 service nova] Acquired lock "refresh_cache-df69e966-4580-42c5-8c88-53008807c1d0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.692926] env[62521]: DEBUG nova.network.neutron [req-23f13881-5a5b-4e02-b58f-a23740e58f76 req-bbc996de-5d18-42e9-ae0c-f9d290ad36d3 service nova] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Refreshing network info cache for port 2fb2566d-d73e-4378-b78e-29be75da5f59 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 567.008541] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52da92fa-4f6a-9404-ee61-69fa39c9dea5, 'name': SearchDatastore_Task, 'duration_secs': 0.030532} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.009478] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d972c4c4-a04b-4ff5-86e2-3d492ed1a110 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.017106] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 567.017106] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520a9910-2b33-7cd4-9c6b-9c1954452c87" [ 567.017106] env[62521]: _type = "Task" [ 567.017106] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.027593] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520a9910-2b33-7cd4-9c6b-9c1954452c87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.143847] env[62521]: DEBUG nova.scheduler.client.report [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.207324] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Successfully created port: 8842b0e6-a852-40b2-8c55-3dcb120fd6d3 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 567.313438] env[62521]: DEBUG nova.network.neutron [req-23f13881-5a5b-4e02-b58f-a23740e58f76 req-bbc996de-5d18-42e9-ae0c-f9d290ad36d3 service nova] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.534024] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520a9910-2b33-7cd4-9c6b-9c1954452c87, 'name': SearchDatastore_Task, 'duration_secs': 0.008835} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.534024] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.534024] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] e976d2d2-aa69-473b-9dfc-e616b0132754/e976d2d2-aa69-473b-9dfc-e616b0132754.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 567.534024] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e989735a-9012-4224-8a56-99f1c5f5ffd6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.541832] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 567.541832] env[62521]: value = "task-1318361" [ 567.541832] env[62521]: _type = "Task" [ 567.541832] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.543587] env[62521]: DEBUG nova.network.neutron [req-23f13881-5a5b-4e02-b58f-a23740e58f76 req-bbc996de-5d18-42e9-ae0c-f9d290ad36d3 service nova] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.556173] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318361, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.650513] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.650513] env[62521]: ERROR nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 05241c37-a746-44c2-ac0d-ac45f086c8d6, please check neutron logs for more information. [ 567.650513] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Traceback (most recent call last): [ 567.650513] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 567.650513] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] self.driver.spawn(context, instance, image_meta, [ 567.650513] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 567.650513] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.650513] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.650513] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] vm_ref = self.build_virtual_machine(instance, [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] for vif in network_info: [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] return self._sync_wrapper(fn, *args, **kwargs) [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] self.wait() [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] self[:] = self._gt.wait() [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] return self._exit_event.wait() [ 567.650827] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] result = hub.switch() [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] return self.greenlet.switch() [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] result = function(*args, **kwargs) [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] return func(*args, **kwargs) [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] raise e [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] nwinfo = self.network_api.allocate_for_instance( [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 567.651212] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] created_port_ids = self._update_ports_for_instance( [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] with excutils.save_and_reraise_exception(): [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] self.force_reraise() [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] raise self.value [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] updated_port = self._update_port( [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] _ensure_no_port_binding_failure(port) [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.651596] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] raise exception.PortBindingFailed(port_id=port['id']) [ 567.651950] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] nova.exception.PortBindingFailed: Binding failed for port 05241c37-a746-44c2-ac0d-ac45f086c8d6, please check neutron logs for more information. [ 567.651950] env[62521]: ERROR nova.compute.manager [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] [ 567.651950] env[62521]: DEBUG nova.compute.utils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Binding failed for port 05241c37-a746-44c2-ac0d-ac45f086c8d6, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 567.652065] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.165s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.657463] env[62521]: DEBUG nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Build of instance 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a was re-scheduled: Binding failed for port 05241c37-a746-44c2-ac0d-ac45f086c8d6, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 567.657463] env[62521]: DEBUG nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 567.657463] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquiring lock "refresh_cache-3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.657463] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquired lock "refresh_cache-3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.658218] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 567.693573] env[62521]: DEBUG nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 567.732735] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 567.732735] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 567.732735] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 567.733626] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 567.733626] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 567.733626] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 567.733841] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 567.733883] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 567.734655] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 567.734655] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 567.734655] env[62521]: DEBUG nova.virt.hardware [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 567.737107] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3f557c-0dca-4248-bd81-423a0301d1b3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.746305] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2b9f62-d23f-4cef-b9f2-43f4193ada5d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.052689] env[62521]: DEBUG oslo_concurrency.lockutils [req-23f13881-5a5b-4e02-b58f-a23740e58f76 req-bbc996de-5d18-42e9-ae0c-f9d290ad36d3 service nova] Releasing lock "refresh_cache-df69e966-4580-42c5-8c88-53008807c1d0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.052932] env[62521]: DEBUG nova.compute.manager [req-23f13881-5a5b-4e02-b58f-a23740e58f76 req-bbc996de-5d18-42e9-ae0c-f9d290ad36d3 service nova] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Received event network-vif-deleted-2fb2566d-d73e-4378-b78e-29be75da5f59 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.053293] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318361, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.211249] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.301223] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.499273] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682efae5-5857-40fb-8eeb-65aa19f8827e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.508716] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898df327-bd7d-47f5-8eb1-ffce5e6555b0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.553294] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144cb4c1-eaff-4926-837a-09b1031d0f20 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.561839] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318361, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.626483} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.564170] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] e976d2d2-aa69-473b-9dfc-e616b0132754/e976d2d2-aa69-473b-9dfc-e616b0132754.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 568.564399] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 568.564685] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ade0d00-bb52-4a05-996a-45827be377b6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.567498] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7880952e-42eb-44e2-8e9a-19df9a7e5725 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.582393] env[62521]: DEBUG nova.compute.provider_tree [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.584962] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 568.584962] env[62521]: value = "task-1318362" [ 568.584962] env[62521]: _type = "Task" [ 568.584962] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.592759] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318362, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.688079] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Acquiring lock "e944e6e5-7bc0-4efb-a811-38221734ec4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.688425] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Lock "e944e6e5-7bc0-4efb-a811-38221734ec4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.810229] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Releasing lock "refresh_cache-3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.810478] env[62521]: DEBUG nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 568.810657] env[62521]: DEBUG nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 568.810814] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 568.838615] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.086747] env[62521]: DEBUG nova.scheduler.client.report [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.103788] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318362, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073714} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.103788] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 569.104739] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e62630-d9b6-47a4-b4bc-7e5a51225815 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.127735] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] e976d2d2-aa69-473b-9dfc-e616b0132754/e976d2d2-aa69-473b-9dfc-e616b0132754.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 569.128380] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24a274fb-88c9-43b1-a4b8-b4d38a73947b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.160603] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 569.160603] env[62521]: value = "task-1318363" [ 569.160603] env[62521]: _type = "Task" [ 569.160603] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.171210] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318363, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.344114] env[62521]: DEBUG nova.network.neutron [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.598792] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.947s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.599492] env[62521]: ERROR nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9f25c76e-cf74-4efb-9207-d77eeef09e72, please check neutron logs for more information. [ 569.599492] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Traceback (most recent call last): [ 569.599492] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.599492] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] self.driver.spawn(context, instance, image_meta, [ 569.599492] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 569.599492] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.599492] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.599492] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] vm_ref = self.build_virtual_machine(instance, [ 569.599492] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.599492] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.599492] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] for vif in network_info: [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] return self._sync_wrapper(fn, *args, **kwargs) [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] self.wait() [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] self[:] = self._gt.wait() [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] return self._exit_event.wait() [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] result = hub.switch() [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.601956] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] return self.greenlet.switch() [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] result = function(*args, **kwargs) [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] return func(*args, **kwargs) [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] raise e [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] nwinfo = self.network_api.allocate_for_instance( [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] created_port_ids = self._update_ports_for_instance( [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] with excutils.save_and_reraise_exception(): [ 569.602656] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] self.force_reraise() [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] raise self.value [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] updated_port = self._update_port( [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] _ensure_no_port_binding_failure(port) [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] raise exception.PortBindingFailed(port_id=port['id']) [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] nova.exception.PortBindingFailed: Binding failed for port 9f25c76e-cf74-4efb-9207-d77eeef09e72, please check neutron logs for more information. [ 569.603169] env[62521]: ERROR nova.compute.manager [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] [ 569.604325] env[62521]: DEBUG nova.compute.utils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Binding failed for port 9f25c76e-cf74-4efb-9207-d77eeef09e72, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 569.604325] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.212s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.604506] env[62521]: DEBUG nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Build of instance bc94294f-206a-4f54-87af-2248c5ed1ba4 was re-scheduled: Binding failed for port 9f25c76e-cf74-4efb-9207-d77eeef09e72, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 569.604972] env[62521]: DEBUG nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 569.605243] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "refresh_cache-bc94294f-206a-4f54-87af-2248c5ed1ba4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.605390] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "refresh_cache-bc94294f-206a-4f54-87af-2248c5ed1ba4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.605547] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 569.655794] env[62521]: INFO nova.compute.manager [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Rebuilding instance [ 569.670330] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318363, 'name': ReconfigVM_Task, 'duration_secs': 0.475977} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.670601] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Reconfigured VM instance instance-0000000a to attach disk [datastore1] e976d2d2-aa69-473b-9dfc-e616b0132754/e976d2d2-aa69-473b-9dfc-e616b0132754.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 569.671186] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc3d6a4c-b4e6-407a-9006-f1e63d67aa84 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.678401] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 569.678401] env[62521]: value = "task-1318365" [ 569.678401] env[62521]: _type = "Task" [ 569.678401] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.700357] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318365, 'name': Rename_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.722651] env[62521]: DEBUG nova.compute.manager [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 569.724554] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02517c36-a967-432d-8fa7-5b155be3cdb9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.825110] env[62521]: ERROR nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8842b0e6-a852-40b2-8c55-3dcb120fd6d3, please check neutron logs for more information. [ 569.825110] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 569.825110] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.825110] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 569.825110] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.825110] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 569.825110] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.825110] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 569.825110] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.825110] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 569.825110] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.825110] env[62521]: ERROR nova.compute.manager raise self.value [ 569.825110] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.825110] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 569.825110] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.825110] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 569.827131] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.827131] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 569.827131] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8842b0e6-a852-40b2-8c55-3dcb120fd6d3, please check neutron logs for more information. [ 569.827131] env[62521]: ERROR nova.compute.manager [ 569.827131] env[62521]: Traceback (most recent call last): [ 569.827131] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 569.827131] env[62521]: listener.cb(fileno) [ 569.827131] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.827131] env[62521]: result = function(*args, **kwargs) [ 569.827131] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 569.827131] env[62521]: return func(*args, **kwargs) [ 569.827131] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.827131] env[62521]: raise e [ 569.827131] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.827131] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 569.827131] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.827131] env[62521]: created_port_ids = self._update_ports_for_instance( [ 569.827131] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.827131] env[62521]: with excutils.save_and_reraise_exception(): [ 569.827131] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.827131] env[62521]: self.force_reraise() [ 569.827131] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.827131] env[62521]: raise self.value [ 569.827131] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.827131] env[62521]: updated_port = self._update_port( [ 569.827131] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.827131] env[62521]: _ensure_no_port_binding_failure(port) [ 569.827131] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.827131] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 569.829144] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 8842b0e6-a852-40b2-8c55-3dcb120fd6d3, please check neutron logs for more information. [ 569.829144] env[62521]: Removing descriptor: 15 [ 569.829144] env[62521]: ERROR nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8842b0e6-a852-40b2-8c55-3dcb120fd6d3, please check neutron logs for more information. [ 569.829144] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] Traceback (most recent call last): [ 569.829144] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 569.829144] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] yield resources [ 569.829144] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.829144] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] self.driver.spawn(context, instance, image_meta, [ 569.829144] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 569.829144] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.829144] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.829144] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] vm_ref = self.build_virtual_machine(instance, [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] for vif in network_info: [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] return self._sync_wrapper(fn, *args, **kwargs) [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] self.wait() [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] self[:] = self._gt.wait() [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] return self._exit_event.wait() [ 569.829723] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] result = hub.switch() [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] return self.greenlet.switch() [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] result = function(*args, **kwargs) [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] return func(*args, **kwargs) [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] raise e [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] nwinfo = self.network_api.allocate_for_instance( [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.830300] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] created_port_ids = self._update_ports_for_instance( [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] with excutils.save_and_reraise_exception(): [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] self.force_reraise() [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] raise self.value [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] updated_port = self._update_port( [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] _ensure_no_port_binding_failure(port) [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.834502] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] raise exception.PortBindingFailed(port_id=port['id']) [ 569.835234] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] nova.exception.PortBindingFailed: Binding failed for port 8842b0e6-a852-40b2-8c55-3dcb120fd6d3, please check neutron logs for more information. [ 569.835234] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] [ 569.835234] env[62521]: INFO nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Terminating instance [ 569.835234] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Acquiring lock "refresh_cache-da34c648-4074-494e-a569-e8e839207a70" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.835234] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Acquired lock "refresh_cache-da34c648-4074-494e-a569-e8e839207a70" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.835234] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 569.845245] env[62521]: INFO nova.compute.manager [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a] Took 1.03 seconds to deallocate network for instance. [ 570.152049] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.201768] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318365, 'name': Rename_Task, 'duration_secs': 0.188526} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.204365] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 570.204465] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f15f1c1c-6af0-488f-bc2c-1c5b9241c7cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.213821] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 570.213821] env[62521]: value = "task-1318366" [ 570.213821] env[62521]: _type = "Task" [ 570.213821] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.225421] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318366, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.240966] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 570.243766] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-756f6cb1-a9b7-4ea9-ac7c-f9d43338253e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.253963] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for the task: (returnval){ [ 570.253963] env[62521]: value = "task-1318367" [ 570.253963] env[62521]: _type = "Task" [ 570.253963] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.270639] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318367, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.380384] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.384577] env[62521]: DEBUG nova.compute.manager [req-cc0fef98-21ad-4eb2-8b5e-97b1f4582a3d req-b01aa53a-2f40-4400-b1c7-73a4789380cd service nova] [instance: da34c648-4074-494e-a569-e8e839207a70] Received event network-changed-8842b0e6-a852-40b2-8c55-3dcb120fd6d3 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 570.384916] env[62521]: DEBUG nova.compute.manager [req-cc0fef98-21ad-4eb2-8b5e-97b1f4582a3d req-b01aa53a-2f40-4400-b1c7-73a4789380cd service nova] [instance: da34c648-4074-494e-a569-e8e839207a70] Refreshing instance network info cache due to event network-changed-8842b0e6-a852-40b2-8c55-3dcb120fd6d3. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 570.385150] env[62521]: DEBUG oslo_concurrency.lockutils [req-cc0fef98-21ad-4eb2-8b5e-97b1f4582a3d req-b01aa53a-2f40-4400-b1c7-73a4789380cd service nova] Acquiring lock "refresh_cache-da34c648-4074-494e-a569-e8e839207a70" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.416308] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.485448] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5551940-69b7-4f5e-88e6-d930f794a8ac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.494445] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5f1bd7-fa91-44a1-ad25-842966519f89 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.546179] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b52fc9a-cd60-4318-85ef-0e52e2191021 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.554564] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9176e04a-36b4-4c16-9f1d-a7830a418558 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.570384] env[62521]: DEBUG nova.compute.provider_tree [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.634182] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.730388] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318366, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.766234] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318367, 'name': PowerOffVM_Task, 'duration_secs': 0.119423} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.767067] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 570.767067] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 570.768367] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33beaf5e-b3c5-4f7a-87de-63c3276e44d0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.784024] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 570.785635] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2264016a-d501-47db-a378-a3d2df75b30d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.810194] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 570.810547] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 570.810854] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Deleting the datastore file [datastore2] 4f3a0a33-43c0-42fb-8660-198ff4a8529a {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 570.812407] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94044cb3-cad6-48b2-bb05-e71ef3fedfbc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.819739] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for the task: (returnval){ [ 570.819739] env[62521]: value = "task-1318369" [ 570.819739] env[62521]: _type = "Task" [ 570.819739] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.828731] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318369, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.879958] env[62521]: INFO nova.scheduler.client.report [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Deleted allocations for instance 3b674a12-d3fb-45f6-94ed-5c2ac2452f7a [ 570.919679] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "refresh_cache-bc94294f-206a-4f54-87af-2248c5ed1ba4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.920077] env[62521]: DEBUG nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 570.920309] env[62521]: DEBUG nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.920494] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 570.948196] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.076543] env[62521]: DEBUG nova.scheduler.client.report [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.087175] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Acquiring lock "fb33026e-b607-477f-b75f-779db7c5c34f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.087175] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Lock "fb33026e-b607-477f-b75f-779db7c5c34f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.138317] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Releasing lock "refresh_cache-da34c648-4074-494e-a569-e8e839207a70" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.138317] env[62521]: DEBUG nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 571.138317] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 571.138427] env[62521]: DEBUG oslo_concurrency.lockutils [req-cc0fef98-21ad-4eb2-8b5e-97b1f4582a3d req-b01aa53a-2f40-4400-b1c7-73a4789380cd service nova] Acquired lock "refresh_cache-da34c648-4074-494e-a569-e8e839207a70" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.138677] env[62521]: DEBUG nova.network.neutron [req-cc0fef98-21ad-4eb2-8b5e-97b1f4582a3d req-b01aa53a-2f40-4400-b1c7-73a4789380cd service nova] [instance: da34c648-4074-494e-a569-e8e839207a70] Refreshing network info cache for port 8842b0e6-a852-40b2-8c55-3dcb120fd6d3 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 571.140439] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-172a6ca6-d40b-4539-b7ba-3cd092839b9d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.150326] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e7b86b-eaf4-4168-9238-b967307cc7d9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.176093] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance da34c648-4074-494e-a569-e8e839207a70 could not be found. [ 571.176342] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 571.176525] env[62521]: INFO nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Took 0.04 seconds to destroy the instance on the hypervisor. [ 571.176773] env[62521]: DEBUG oslo.service.loopingcall [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.177888] env[62521]: DEBUG nova.compute.manager [-] [instance: da34c648-4074-494e-a569-e8e839207a70] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.177984] env[62521]: DEBUG nova.network.neutron [-] [instance: da34c648-4074-494e-a569-e8e839207a70] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 571.203535] env[62521]: DEBUG nova.network.neutron [-] [instance: da34c648-4074-494e-a569-e8e839207a70] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.226648] env[62521]: DEBUG oslo_vmware.api [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318366, 'name': PowerOnVM_Task, 'duration_secs': 0.750175} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.226869] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 571.231461] env[62521]: INFO nova.compute.manager [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Took 5.91 seconds to spawn the instance on the hypervisor. [ 571.231775] env[62521]: DEBUG nova.compute.manager [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 571.235016] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab23e67-cd98-4235-babb-d7b58d493b1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.329637] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318369, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165924} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.329975] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 571.330099] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 571.330271] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 571.390567] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3661c757-6795-48fc-bbcc-647ea408ef22 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "3b674a12-d3fb-45f6-94ed-5c2ac2452f7a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.294s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.452371] env[62521]: DEBUG nova.network.neutron [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.583732] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.982s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.584431] env[62521]: ERROR nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bfd3f27e-25d5-445e-8451-d2511d8c40f3, please check neutron logs for more information. [ 571.584431] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Traceback (most recent call last): [ 571.584431] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 571.584431] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] self.driver.spawn(context, instance, image_meta, [ 571.584431] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 571.584431] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.584431] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.584431] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] vm_ref = self.build_virtual_machine(instance, [ 571.584431] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.584431] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.584431] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] for vif in network_info: [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] return self._sync_wrapper(fn, *args, **kwargs) [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] self.wait() [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] self[:] = self._gt.wait() [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] return self._exit_event.wait() [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] result = hub.switch() [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.584785] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] return self.greenlet.switch() [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] result = function(*args, **kwargs) [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] return func(*args, **kwargs) [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] raise e [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] nwinfo = self.network_api.allocate_for_instance( [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] created_port_ids = self._update_ports_for_instance( [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] with excutils.save_and_reraise_exception(): [ 571.585244] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] self.force_reraise() [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] raise self.value [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] updated_port = self._update_port( [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] _ensure_no_port_binding_failure(port) [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] raise exception.PortBindingFailed(port_id=port['id']) [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] nova.exception.PortBindingFailed: Binding failed for port bfd3f27e-25d5-445e-8451-d2511d8c40f3, please check neutron logs for more information. [ 571.585709] env[62521]: ERROR nova.compute.manager [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] [ 571.586074] env[62521]: DEBUG nova.compute.utils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Binding failed for port bfd3f27e-25d5-445e-8451-d2511d8c40f3, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 571.587135] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.055s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.595658] env[62521]: DEBUG nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Build of instance 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745 was re-scheduled: Binding failed for port bfd3f27e-25d5-445e-8451-d2511d8c40f3, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 571.596304] env[62521]: DEBUG nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 571.596554] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Acquiring lock "refresh_cache-0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.598264] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Acquired lock "refresh_cache-0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.598264] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 571.672714] env[62521]: DEBUG nova.network.neutron [req-cc0fef98-21ad-4eb2-8b5e-97b1f4582a3d req-b01aa53a-2f40-4400-b1c7-73a4789380cd service nova] [instance: da34c648-4074-494e-a569-e8e839207a70] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.707374] env[62521]: DEBUG nova.network.neutron [-] [instance: da34c648-4074-494e-a569-e8e839207a70] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.774742] env[62521]: INFO nova.compute.manager [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Took 21.34 seconds to build instance. [ 571.893714] env[62521]: DEBUG nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 571.922767] env[62521]: DEBUG nova.network.neutron [req-cc0fef98-21ad-4eb2-8b5e-97b1f4582a3d req-b01aa53a-2f40-4400-b1c7-73a4789380cd service nova] [instance: da34c648-4074-494e-a569-e8e839207a70] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.956255] env[62521]: INFO nova.compute.manager [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: bc94294f-206a-4f54-87af-2248c5ed1ba4] Took 1.04 seconds to deallocate network for instance. [ 572.059787] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquiring lock "d9571cc5-0555-4867-a8d3-a2281d512305" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.059935] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "d9571cc5-0555-4867-a8d3-a2281d512305" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.133616] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.210723] env[62521]: INFO nova.compute.manager [-] [instance: da34c648-4074-494e-a569-e8e839207a70] Took 1.03 seconds to deallocate network for instance. [ 572.218027] env[62521]: DEBUG nova.compute.claims [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 572.222365] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.223754] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.278342] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f859ebe9-4798-4db4-ab37-aad541433040 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lock "e976d2d2-aa69-473b-9dfc-e616b0132754" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.855s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.372009] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.372009] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.372009] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.372009] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.372489] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.372489] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.372489] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.372606] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.372990] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.373982] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.375258] env[62521]: DEBUG nova.virt.hardware [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.375967] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8ccb0e-024e-4a18-9d71-aa0627c56ae4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.391623] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d91730-ddd2-4351-8baf-3275de029f29 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.415737] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 572.421857] env[62521]: DEBUG oslo.service.loopingcall [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 572.427173] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 572.427643] env[62521]: DEBUG oslo_concurrency.lockutils [req-cc0fef98-21ad-4eb2-8b5e-97b1f4582a3d req-b01aa53a-2f40-4400-b1c7-73a4789380cd service nova] Releasing lock "refresh_cache-da34c648-4074-494e-a569-e8e839207a70" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.428881] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-757f04b5-d148-4d51-aa7d-39259afb29ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.441197] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.449680] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 572.449680] env[62521]: value = "task-1318371" [ 572.449680] env[62521]: _type = "Task" [ 572.449680] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.465045] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318371, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.480908] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067f635d-4f42-4a95-a59f-1c3e3bcbb966 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.488444] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8602de9e-e7d3-46f6-8ad1-c5216f2f6187 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.525973] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8a2549-84a3-406f-b603-99f7622693e9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.535309] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233be417-6580-4d43-993a-24ba7ac9684a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.550967] env[62521]: DEBUG nova.compute.provider_tree [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.727580] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Releasing lock "refresh_cache-0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.727790] env[62521]: DEBUG nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 572.728058] env[62521]: DEBUG nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 572.728397] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 572.751036] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.782260] env[62521]: DEBUG nova.compute.manager [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 572.962842] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318371, 'name': CreateVM_Task, 'duration_secs': 0.37772} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.962972] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 572.965637] env[62521]: DEBUG oslo_concurrency.lockutils [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.965919] env[62521]: DEBUG oslo_concurrency.lockutils [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.970773] env[62521]: DEBUG oslo_concurrency.lockutils [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 572.976109] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c96a7e62-2fb5-48a3-a8b3-863d96944a84 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.983199] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for the task: (returnval){ [ 572.983199] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f3dba2-42bd-a692-c586-2f5e98df26d8" [ 572.983199] env[62521]: _type = "Task" [ 572.983199] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.994411] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f3dba2-42bd-a692-c586-2f5e98df26d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.018734] env[62521]: INFO nova.scheduler.client.report [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleted allocations for instance bc94294f-206a-4f54-87af-2248c5ed1ba4 [ 573.054066] env[62521]: DEBUG nova.scheduler.client.report [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.254420] env[62521]: DEBUG nova.network.neutron [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.306220] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.447081] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquiring lock "fe69caab-dd71-4b01-a121-2591111b193d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.447345] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "fe69caab-dd71-4b01-a121-2591111b193d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.498301] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f3dba2-42bd-a692-c586-2f5e98df26d8, 'name': SearchDatastore_Task, 'duration_secs': 0.013385} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.498301] env[62521]: DEBUG oslo_concurrency.lockutils [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.498301] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 573.498301] env[62521]: DEBUG oslo_concurrency.lockutils [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.498524] env[62521]: DEBUG oslo_concurrency.lockutils [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.498524] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 573.498524] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7c785662-9cff-4008-a4b1-d632e8e3d513 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.506172] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 573.507271] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 573.507271] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df47505a-4c05-4325-93ae-e8cf546bc131 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.514398] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for the task: (returnval){ [ 573.514398] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b82995-7241-8378-5a07-81cef214e16f" [ 573.514398] env[62521]: _type = "Task" [ 573.514398] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.523431] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b82995-7241-8378-5a07-81cef214e16f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.538868] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b4ed454a-8c79-4832-8e40-ffbcacfe3670 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "bc94294f-206a-4f54-87af-2248c5ed1ba4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.385s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.563240] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.563240] env[62521]: ERROR nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f943804f-fca5-4c9b-96da-fe5bd2df773d, please check neutron logs for more information. [ 573.563240] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Traceback (most recent call last): [ 573.563240] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.563240] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] self.driver.spawn(context, instance, image_meta, [ 573.563240] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 573.563240] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.563240] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.563240] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] vm_ref = self.build_virtual_machine(instance, [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] for vif in network_info: [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] return self._sync_wrapper(fn, *args, **kwargs) [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] self.wait() [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] self[:] = self._gt.wait() [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] return self._exit_event.wait() [ 573.563801] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] result = hub.switch() [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] return self.greenlet.switch() [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] result = function(*args, **kwargs) [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] return func(*args, **kwargs) [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] raise e [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] nwinfo = self.network_api.allocate_for_instance( [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.564468] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] created_port_ids = self._update_ports_for_instance( [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] with excutils.save_and_reraise_exception(): [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] self.force_reraise() [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] raise self.value [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] updated_port = self._update_port( [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] _ensure_no_port_binding_failure(port) [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.564838] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] raise exception.PortBindingFailed(port_id=port['id']) [ 573.565309] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] nova.exception.PortBindingFailed: Binding failed for port f943804f-fca5-4c9b-96da-fe5bd2df773d, please check neutron logs for more information. [ 573.565309] env[62521]: ERROR nova.compute.manager [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] [ 573.565309] env[62521]: DEBUG nova.compute.utils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Binding failed for port f943804f-fca5-4c9b-96da-fe5bd2df773d, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.566285] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.632s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.571021] env[62521]: DEBUG nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Build of instance 71c6b4ec-60a5-4c8d-a529-a841e1b36a53 was re-scheduled: Binding failed for port f943804f-fca5-4c9b-96da-fe5bd2df773d, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.573030] env[62521]: DEBUG nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 573.573030] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Acquiring lock "refresh_cache-71c6b4ec-60a5-4c8d-a529-a841e1b36a53" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.573030] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Acquired lock "refresh_cache-71c6b4ec-60a5-4c8d-a529-a841e1b36a53" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.573030] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 573.759757] env[62521]: INFO nova.compute.manager [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] [instance: 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745] Took 1.03 seconds to deallocate network for instance. [ 573.771495] env[62521]: DEBUG nova.compute.manager [req-4d4cb60c-3feb-4aeb-9457-0ede4120572d req-24c95bb1-a1d5-4c2a-a411-25b59abff37c service nova] [instance: da34c648-4074-494e-a569-e8e839207a70] Received event network-vif-deleted-8842b0e6-a852-40b2-8c55-3dcb120fd6d3 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 573.880834] env[62521]: DEBUG nova.compute.manager [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 573.882166] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cac68e6-3de1-4974-b1cc-738022e32676 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.029714] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b82995-7241-8378-5a07-81cef214e16f, 'name': SearchDatastore_Task, 'duration_secs': 0.009083} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.030518] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99829347-ef6c-4cf8-87ba-5e2fe8219823 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.036688] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for the task: (returnval){ [ 574.036688] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]528cd662-26f7-99e7-7fe2-66ad080e895b" [ 574.036688] env[62521]: _type = "Task" [ 574.036688] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.045248] env[62521]: DEBUG nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 574.047629] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]528cd662-26f7-99e7-7fe2-66ad080e895b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.107825] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.212458] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Acquiring lock "edd89a57-f026-4a3e-8b6b-111cc2321d9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.212458] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Lock "edd89a57-f026-4a3e-8b6b-111cc2321d9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.234408] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.395874] env[62521]: INFO nova.compute.manager [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] instance snapshotting [ 574.395874] env[62521]: DEBUG nova.objects.instance [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lazy-loading 'flavor' on Instance uuid e976d2d2-aa69-473b-9dfc-e616b0132754 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 574.547417] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]528cd662-26f7-99e7-7fe2-66ad080e895b, 'name': SearchDatastore_Task, 'duration_secs': 0.008698} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.547713] env[62521]: DEBUG oslo_concurrency.lockutils [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.547929] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 574.548422] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30a7e025-359c-4701-849c-e72a309e52db {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.560423] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for the task: (returnval){ [ 574.560423] env[62521]: value = "task-1318372" [ 574.560423] env[62521]: _type = "Task" [ 574.560423] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.569336] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318372, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.581090] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.613191] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 4f3a0a33-43c0-42fb-8660-198ff4a8529a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 574.738957] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Releasing lock "refresh_cache-71c6b4ec-60a5-4c8d-a529-a841e1b36a53" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.739915] env[62521]: DEBUG nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.739915] env[62521]: DEBUG nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.739915] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.762913] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.821260] env[62521]: INFO nova.scheduler.client.report [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Deleted allocations for instance 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745 [ 574.900676] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52de3f3c-6ed8-45c1-8ae7-df0ef0168360 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.921610] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c96174-f480-453e-ad17-4760c34dee69 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.071967] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318372, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.116905] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 0ecd5e0c-e61d-45ff-ba4e-70ff83b26745 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 575.267318] env[62521]: DEBUG nova.network.neutron [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.335626] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9fa19c22-9b59-4389-ad3d-a7d95d6e117d tempest-ServerExternalEventsTest-768488011 tempest-ServerExternalEventsTest-768488011-project-member] Lock "0ecd5e0c-e61d-45ff-ba4e-70ff83b26745" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.887s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.436022] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 575.436022] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b7d9d5a4-3ff5-43ae-aa82-20c61cd55753 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.444018] env[62521]: DEBUG oslo_vmware.api [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 575.444018] env[62521]: value = "task-1318373" [ 575.444018] env[62521]: _type = "Task" [ 575.444018] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.455012] env[62521]: DEBUG oslo_vmware.api [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318373, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.573365] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318372, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574253} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.573365] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 575.574240] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 575.574353] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04a71b90-0057-4f43-aada-b51277a6a9a5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.581719] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for the task: (returnval){ [ 575.581719] env[62521]: value = "task-1318375" [ 575.581719] env[62521]: _type = "Task" [ 575.581719] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.591578] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318375, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.627326] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 71c6b4ec-60a5-4c8d-a529-a841e1b36a53 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 575.627326] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance c3a43b3a-cb59-4431-a5d2-4c96db16337f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 575.627326] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance dba06415-d503-4420-825e-76afc59f6153 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 575.627326] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance df69e966-4580-42c5-8c88-53008807c1d0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 575.627625] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance e976d2d2-aa69-473b-9dfc-e616b0132754 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 575.627625] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance da34c648-4074-494e-a569-e8e839207a70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 575.778412] env[62521]: INFO nova.compute.manager [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] [instance: 71c6b4ec-60a5-4c8d-a529-a841e1b36a53] Took 1.04 seconds to deallocate network for instance. [ 575.839187] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 575.917744] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquiring lock "e976d2d2-aa69-473b-9dfc-e616b0132754" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.920677] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lock "e976d2d2-aa69-473b-9dfc-e616b0132754" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.920677] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquiring lock "e976d2d2-aa69-473b-9dfc-e616b0132754-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.920677] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lock "e976d2d2-aa69-473b-9dfc-e616b0132754-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.920677] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lock "e976d2d2-aa69-473b-9dfc-e616b0132754-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.921757] env[62521]: INFO nova.compute.manager [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Terminating instance [ 575.925817] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquiring lock "refresh_cache-e976d2d2-aa69-473b-9dfc-e616b0132754" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.925892] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquired lock "refresh_cache-e976d2d2-aa69-473b-9dfc-e616b0132754" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.926392] env[62521]: DEBUG nova.network.neutron [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 575.953854] env[62521]: DEBUG oslo_vmware.api [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318373, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.094375] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318375, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064645} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.094699] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 576.095588] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0a5119-cf0d-4fe1-a350-d8402691fb61 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.124278] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 576.124871] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da3e6b80-e0d8-4065-a26f-afab40daafe9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.142880] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 20b01df0-4ced-4668-a786-046251a0c7ed has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 576.152195] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for the task: (returnval){ [ 576.152195] env[62521]: value = "task-1318376" [ 576.152195] env[62521]: _type = "Task" [ 576.152195] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.162237] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318376, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.370986] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.453999] env[62521]: DEBUG oslo_vmware.api [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318373, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.458442] env[62521]: DEBUG nova.network.neutron [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.647084] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 576.665024] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318376, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.741133] env[62521]: DEBUG nova.network.neutron [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.829661] env[62521]: INFO nova.scheduler.client.report [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Deleted allocations for instance 71c6b4ec-60a5-4c8d-a529-a841e1b36a53 [ 576.961213] env[62521]: DEBUG oslo_vmware.api [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318373, 'name': CreateSnapshot_Task, 'duration_secs': 1.32326} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.961551] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 576.962652] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8feda579-17bc-4444-ad58-aef674797777 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.152679] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 7653f469-5ce5-4d74-9a3f-55a6adb3a255 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 577.171210] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318376, 'name': ReconfigVM_Task, 'duration_secs': 0.808322} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.174674] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Reconfigured VM instance instance-00000001 to attach disk [datastore2] 4f3a0a33-43c0-42fb-8660-198ff4a8529a/4f3a0a33-43c0-42fb-8660-198ff4a8529a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 577.175370] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8c5265b-fba6-4f37-97e6-180a0b057634 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.185512] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for the task: (returnval){ [ 577.185512] env[62521]: value = "task-1318377" [ 577.185512] env[62521]: _type = "Task" [ 577.185512] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.194261] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318377, 'name': Rename_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.243487] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Releasing lock "refresh_cache-e976d2d2-aa69-473b-9dfc-e616b0132754" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.243995] env[62521]: DEBUG nova.compute.manager [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 577.244216] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 577.245448] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e60de0-94d4-4b71-b35d-788f78d46235 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.256833] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 577.257598] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf7fe2b0-853b-47ba-9895-200b08880c1c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.265392] env[62521]: DEBUG oslo_vmware.api [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 577.265392] env[62521]: value = "task-1318378" [ 577.265392] env[62521]: _type = "Task" [ 577.265392] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.277649] env[62521]: DEBUG oslo_vmware.api [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318378, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.344338] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b74a208f-8c5a-470f-afba-bc3a03ccf6ad tempest-ServersTestJSON-438512769 tempest-ServersTestJSON-438512769-project-member] Lock "71c6b4ec-60a5-4c8d-a529-a841e1b36a53" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.082s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.476489] env[62521]: DEBUG nova.compute.manager [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Instance disappeared during snapshot {{(pid=62521) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 577.657052] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 6296d4ea-d678-4981-bc5a-729e83020fa2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 577.661717] env[62521]: DEBUG nova.compute.manager [None req-4d6296c3-b4d0-4bb9-a822-fc66c9ec3e81 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Found 0 images (rotation: 2) {{(pid=62521) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 577.697191] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318377, 'name': Rename_Task, 'duration_secs': 0.162418} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.697679] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 577.698454] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b096a302-4a46-4861-ad60-63c84693abf8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.705664] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Waiting for the task: (returnval){ [ 577.705664] env[62521]: value = "task-1318379" [ 577.705664] env[62521]: _type = "Task" [ 577.705664] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.718222] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318379, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.775700] env[62521]: DEBUG oslo_vmware.api [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318378, 'name': PowerOffVM_Task, 'duration_secs': 0.156291} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.776426] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 577.776426] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 577.776603] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09e16995-eb45-4f40-b2db-a47c3daf6302 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.807320] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquiring lock "104b64ff-a042-4bee-a5fd-bbbdf288f63f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.807696] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "104b64ff-a042-4bee-a5fd-bbbdf288f63f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.808883] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 577.809263] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 577.809263] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Deleting the datastore file [datastore1] e976d2d2-aa69-473b-9dfc-e616b0132754 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 577.809786] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ed842ea-3d1d-4965-9aa9-47fa3ec706f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.817892] env[62521]: DEBUG oslo_vmware.api [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for the task: (returnval){ [ 577.817892] env[62521]: value = "task-1318381" [ 577.817892] env[62521]: _type = "Task" [ 577.817892] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.830546] env[62521]: DEBUG oslo_vmware.api [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318381, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.848018] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 578.160303] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 4b2fda93-a60d-4050-ae05-7a37ac284917 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.216731] env[62521]: DEBUG oslo_vmware.api [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Task: {'id': task-1318379, 'name': PowerOnVM_Task, 'duration_secs': 0.422042} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.217269] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 578.217380] env[62521]: DEBUG nova.compute.manager [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 578.218189] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e8e150-0ab5-4583-879b-9d5cb8b7307c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.335012] env[62521]: DEBUG oslo_vmware.api [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Task: {'id': task-1318381, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199718} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.335012] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 578.335012] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 578.335012] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 578.335012] env[62521]: INFO nova.compute.manager [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Took 1.09 seconds to destroy the instance on the hypervisor. [ 578.335349] env[62521]: DEBUG oslo.service.loopingcall [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 578.335349] env[62521]: DEBUG nova.compute.manager [-] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.335349] env[62521]: DEBUG nova.network.neutron [-] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 578.358638] env[62521]: DEBUG nova.network.neutron [-] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.368411] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.663600] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 1d25cdca-3c87-47a2-85aa-86f7132297b5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.743219] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Acquiring lock "e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.743219] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Lock "e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.743219] env[62521]: DEBUG oslo_concurrency.lockutils [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.860384] env[62521]: DEBUG nova.network.neutron [-] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.126860] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "215472ea-402c-4eb5-8ad3-b7ef756dfdc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.127050] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "215472ea-402c-4eb5-8ad3-b7ef756dfdc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.168677] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 1c892e78-a863-4561-940d-b44701e97041 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 579.363680] env[62521]: INFO nova.compute.manager [-] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Took 1.03 seconds to deallocate network for instance. [ 579.673873] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 9c2c0e66-5c67-4e0a-9205-4080bda44a66 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 579.781225] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "f7611f03-73bb-4c0c-b787-d3c7f077f5bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.781462] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "f7611f03-73bb-4c0c-b787-d3c7f077f5bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.871342] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.177694] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance da9d2279-79ae-4371-bef5-08ae15f2d9c9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 580.590905] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Acquiring lock "e1ead278-d904-41cc-bdfe-65174e8d6816" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.591118] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Lock "e1ead278-d904-41cc-bdfe-65174e8d6816" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.676404] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "4f3a0a33-43c0-42fb-8660-198ff4a8529a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.676769] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "4f3a0a33-43c0-42fb-8660-198ff4a8529a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.677018] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "4f3a0a33-43c0-42fb-8660-198ff4a8529a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.677217] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "4f3a0a33-43c0-42fb-8660-198ff4a8529a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.677406] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "4f3a0a33-43c0-42fb-8660-198ff4a8529a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.680141] env[62521]: INFO nova.compute.manager [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Terminating instance [ 580.682338] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance e944e6e5-7bc0-4efb-a811-38221734ec4f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 580.690550] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "refresh_cache-4f3a0a33-43c0-42fb-8660-198ff4a8529a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.690619] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquired lock "refresh_cache-4f3a0a33-43c0-42fb-8660-198ff4a8529a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.690813] env[62521]: DEBUG nova.network.neutron [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.193699] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance fb33026e-b607-477f-b75f-779db7c5c34f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 581.233685] env[62521]: DEBUG nova.network.neutron [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.389216] env[62521]: DEBUG nova.network.neutron [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.703228] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance d9571cc5-0555-4867-a8d3-a2281d512305 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 581.892556] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Releasing lock "refresh_cache-4f3a0a33-43c0-42fb-8660-198ff4a8529a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.895539] env[62521]: DEBUG nova.compute.manager [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 581.895539] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 581.896250] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3ecb21-9593-446c-bc08-0aeee2d5ce21 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.911159] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 581.911159] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0259a235-a363-40af-b3ca-08d18aac2e2b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.915288] env[62521]: DEBUG oslo_vmware.api [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 581.915288] env[62521]: value = "task-1318382" [ 581.915288] env[62521]: _type = "Task" [ 581.915288] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.927019] env[62521]: DEBUG oslo_vmware.api [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.207707] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance fe69caab-dd71-4b01-a121-2591111b193d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 582.425667] env[62521]: DEBUG oslo_vmware.api [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318382, 'name': PowerOffVM_Task, 'duration_secs': 0.134663} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.425933] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 582.426115] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 582.426365] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0f0a8c2-df18-46cd-8254-243d430bfc5a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.454229] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 582.454562] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 582.454731] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleting the datastore file [datastore2] 4f3a0a33-43c0-42fb-8660-198ff4a8529a {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 582.455063] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1169f345-9a25-4aa4-ba67-245b6378c0ee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.461331] env[62521]: DEBUG oslo_vmware.api [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for the task: (returnval){ [ 582.461331] env[62521]: value = "task-1318384" [ 582.461331] env[62521]: _type = "Task" [ 582.461331] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.469956] env[62521]: DEBUG oslo_vmware.api [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318384, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.714676] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance edd89a57-f026-4a3e-8b6b-111cc2321d9d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 582.714676] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 582.714676] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 582.972357] env[62521]: DEBUG oslo_vmware.api [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Task: {'id': task-1318384, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106716} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.972581] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 582.972831] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 582.973091] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 582.973548] env[62521]: INFO nova.compute.manager [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Took 1.08 seconds to destroy the instance on the hypervisor. [ 582.973662] env[62521]: DEBUG oslo.service.loopingcall [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.977721] env[62521]: DEBUG nova.compute.manager [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.977822] env[62521]: DEBUG nova.network.neutron [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.016577] env[62521]: DEBUG nova.network.neutron [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.153848] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16576723-d238-4a21-83ab-9373538e9990 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.165127] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304d3f48-1b88-4ce2-bf52-677fd282d181 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.198395] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781446d7-6c29-4af6-8acb-7850dcd01dcb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.206289] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809d6c1c-90de-41f7-95d2-a79b1c92b0ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.223346] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.517895] env[62521]: DEBUG nova.network.neutron [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.727650] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.025459] env[62521]: INFO nova.compute.manager [-] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Took 1.05 seconds to deallocate network for instance. [ 584.234165] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 584.234316] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.668s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.234608] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.221s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.537485] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.138046] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b245ffe-c136-43f2-9c1a-477c21cc64b4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.148299] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533e51d6-950c-4393-9614-219996f8a01c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.185477] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-431aaf84-bd04-4948-a772-09bd76417bfc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.195232] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e8f58e-8cb8-44e7-b94e-aec70a2ee29f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.211974] env[62521]: DEBUG nova.compute.provider_tree [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.716081] env[62521]: DEBUG nova.scheduler.client.report [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.223325] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.989s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.224075] env[62521]: ERROR nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7e1b61ae-32ea-4647-9046-23f7241687c2, please check neutron logs for more information. [ 586.224075] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Traceback (most recent call last): [ 586.224075] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.224075] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] self.driver.spawn(context, instance, image_meta, [ 586.224075] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 586.224075] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.224075] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.224075] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] vm_ref = self.build_virtual_machine(instance, [ 586.224075] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.224075] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.224075] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] for vif in network_info: [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] return self._sync_wrapper(fn, *args, **kwargs) [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] self.wait() [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] self[:] = self._gt.wait() [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] return self._exit_event.wait() [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] result = hub.switch() [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.224473] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] return self.greenlet.switch() [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] result = function(*args, **kwargs) [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] return func(*args, **kwargs) [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] raise e [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] nwinfo = self.network_api.allocate_for_instance( [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] created_port_ids = self._update_ports_for_instance( [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] with excutils.save_and_reraise_exception(): [ 586.225440] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] self.force_reraise() [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] raise self.value [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] updated_port = self._update_port( [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] _ensure_no_port_binding_failure(port) [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] raise exception.PortBindingFailed(port_id=port['id']) [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] nova.exception.PortBindingFailed: Binding failed for port 7e1b61ae-32ea-4647-9046-23f7241687c2, please check neutron logs for more information. [ 586.225821] env[62521]: ERROR nova.compute.manager [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] [ 586.226191] env[62521]: DEBUG nova.compute.utils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Binding failed for port 7e1b61ae-32ea-4647-9046-23f7241687c2, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 586.226191] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.871s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.229267] env[62521]: DEBUG nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Build of instance c3a43b3a-cb59-4431-a5d2-4c96db16337f was re-scheduled: Binding failed for port 7e1b61ae-32ea-4647-9046-23f7241687c2, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 586.230735] env[62521]: DEBUG nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 586.230735] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Acquiring lock "refresh_cache-c3a43b3a-cb59-4431-a5d2-4c96db16337f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.230735] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Acquired lock "refresh_cache-c3a43b3a-cb59-4431-a5d2-4c96db16337f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.230735] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 586.764380] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.841330] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.125787] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ff3500-040b-4017-858c-66618232dde9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.132754] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79bfae1-5963-48d2-b88d-5d4b725656d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.173696] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b348d5c2-0aee-4d9a-acfd-5d19999c8c0a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.182386] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4c7cd5-0c4b-4bcb-90df-e481bb459418 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.200895] env[62521]: DEBUG nova.compute.provider_tree [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.343245] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Releasing lock "refresh_cache-c3a43b3a-cb59-4431-a5d2-4c96db16337f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.343494] env[62521]: DEBUG nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 587.343604] env[62521]: DEBUG nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.343829] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.361387] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.703700] env[62521]: DEBUG nova.scheduler.client.report [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.867418] env[62521]: DEBUG nova.network.neutron [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.210303] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.211193] env[62521]: ERROR nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d12d3198-135b-4dc6-aafe-d02cca92d79e, please check neutron logs for more information. [ 588.211193] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] Traceback (most recent call last): [ 588.211193] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.211193] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] self.driver.spawn(context, instance, image_meta, [ 588.211193] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 588.211193] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.211193] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.211193] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] vm_ref = self.build_virtual_machine(instance, [ 588.211193] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.211193] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.211193] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] for vif in network_info: [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] return self._sync_wrapper(fn, *args, **kwargs) [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] self.wait() [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] self[:] = self._gt.wait() [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] return self._exit_event.wait() [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] result = hub.switch() [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.211650] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] return self.greenlet.switch() [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] result = function(*args, **kwargs) [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] return func(*args, **kwargs) [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] raise e [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] nwinfo = self.network_api.allocate_for_instance( [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] created_port_ids = self._update_ports_for_instance( [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] with excutils.save_and_reraise_exception(): [ 588.212177] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] self.force_reraise() [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] raise self.value [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] updated_port = self._update_port( [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] _ensure_no_port_binding_failure(port) [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] raise exception.PortBindingFailed(port_id=port['id']) [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] nova.exception.PortBindingFailed: Binding failed for port d12d3198-135b-4dc6-aafe-d02cca92d79e, please check neutron logs for more information. [ 588.212677] env[62521]: ERROR nova.compute.manager [instance: dba06415-d503-4420-825e-76afc59f6153] [ 588.213125] env[62521]: DEBUG nova.compute.utils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Binding failed for port d12d3198-135b-4dc6-aafe-d02cca92d79e, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 588.213125] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.429s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.214539] env[62521]: INFO nova.compute.claims [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.217954] env[62521]: DEBUG nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Build of instance dba06415-d503-4420-825e-76afc59f6153 was re-scheduled: Binding failed for port d12d3198-135b-4dc6-aafe-d02cca92d79e, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 588.217954] env[62521]: DEBUG nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 588.217954] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquiring lock "refresh_cache-dba06415-d503-4420-825e-76afc59f6153" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.218202] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquired lock "refresh_cache-dba06415-d503-4420-825e-76afc59f6153" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.218202] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.371813] env[62521]: INFO nova.compute.manager [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] [instance: c3a43b3a-cb59-4431-a5d2-4c96db16337f] Took 1.03 seconds to deallocate network for instance. [ 588.744848] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.841320] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.199927] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Acquiring lock "166786ee-4b89-4683-831b-9f3c1c57fb08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.200167] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Lock "166786ee-4b89-4683-831b-9f3c1c57fb08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.346458] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Releasing lock "refresh_cache-dba06415-d503-4420-825e-76afc59f6153" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.346458] env[62521]: DEBUG nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 589.346607] env[62521]: DEBUG nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.346773] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.363804] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.401614] env[62521]: INFO nova.scheduler.client.report [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Deleted allocations for instance c3a43b3a-cb59-4431-a5d2-4c96db16337f [ 589.514503] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Acquiring lock "cd3d793a-aa56-4f07-a2fa-c56693548dfb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.514671] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Lock "cd3d793a-aa56-4f07-a2fa-c56693548dfb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.634578] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928cbf10-4571-4c32-9cd3-5e9e0ff72052 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.643059] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc161f7-6d3b-4cf7-8ed8-932149967bc3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.673884] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f130b342-c794-4f7e-b333-04218ee748dd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.681705] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5994588-8734-4f46-9c64-c475946a7799 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.697875] env[62521]: DEBUG nova.compute.provider_tree [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.867555] env[62521]: DEBUG nova.network.neutron [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.912349] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e76ed30-65e3-4532-85ef-bf7258b76152 tempest-TenantUsagesTestJSON-1446589224 tempest-TenantUsagesTestJSON-1446589224-project-member] Lock "c3a43b3a-cb59-4431-a5d2-4c96db16337f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.684s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.204611] env[62521]: DEBUG nova.scheduler.client.report [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.375385] env[62521]: INFO nova.compute.manager [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: dba06415-d503-4420-825e-76afc59f6153] Took 1.03 seconds to deallocate network for instance. [ 590.415078] env[62521]: DEBUG nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 590.712185] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.712761] env[62521]: DEBUG nova.compute.manager [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 590.716059] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.704s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.943283] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.227091] env[62521]: DEBUG nova.compute.utils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.235234] env[62521]: DEBUG nova.compute.manager [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Not allocating networking since 'none' was specified. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 591.414136] env[62521]: INFO nova.scheduler.client.report [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Deleted allocations for instance dba06415-d503-4420-825e-76afc59f6153 [ 591.655455] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4a78e6-56f4-488d-92c0-0fff46ffb3ac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.664082] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eabf46a-275f-498f-aef5-c795532db3ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.698381] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1710e02f-907b-483e-bc5b-d1b806ad6559 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.706916] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e2e516-7c7a-4b86-aaac-ad988ffc7f37 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.726201] env[62521]: DEBUG nova.compute.provider_tree [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.737467] env[62521]: DEBUG nova.compute.manager [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.922762] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99b712f9-7d4f-4355-92a8-7b8f8b56c01d tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "dba06415-d503-4420-825e-76afc59f6153" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.411s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.232018] env[62521]: DEBUG nova.scheduler.client.report [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.426490] env[62521]: DEBUG nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 592.735608] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.019s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.736278] env[62521]: ERROR nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2fb2566d-d73e-4378-b78e-29be75da5f59, please check neutron logs for more information. [ 592.736278] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] Traceback (most recent call last): [ 592.736278] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.736278] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] self.driver.spawn(context, instance, image_meta, [ 592.736278] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 592.736278] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.736278] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.736278] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] vm_ref = self.build_virtual_machine(instance, [ 592.736278] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.736278] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.736278] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] for vif in network_info: [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] return self._sync_wrapper(fn, *args, **kwargs) [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] self.wait() [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] self[:] = self._gt.wait() [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] return self._exit_event.wait() [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] result = hub.switch() [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.736669] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] return self.greenlet.switch() [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] result = function(*args, **kwargs) [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] return func(*args, **kwargs) [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] raise e [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] nwinfo = self.network_api.allocate_for_instance( [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] created_port_ids = self._update_ports_for_instance( [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] with excutils.save_and_reraise_exception(): [ 592.737072] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] self.force_reraise() [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] raise self.value [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] updated_port = self._update_port( [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] _ensure_no_port_binding_failure(port) [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] raise exception.PortBindingFailed(port_id=port['id']) [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] nova.exception.PortBindingFailed: Binding failed for port 2fb2566d-d73e-4378-b78e-29be75da5f59, please check neutron logs for more information. [ 592.737462] env[62521]: ERROR nova.compute.manager [instance: df69e966-4580-42c5-8c88-53008807c1d0] [ 592.737815] env[62521]: DEBUG nova.compute.utils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Binding failed for port 2fb2566d-d73e-4378-b78e-29be75da5f59, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 592.740870] env[62521]: DEBUG nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Build of instance df69e966-4580-42c5-8c88-53008807c1d0 was re-scheduled: Binding failed for port 2fb2566d-d73e-4378-b78e-29be75da5f59, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 592.741373] env[62521]: DEBUG nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 592.741647] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquiring lock "refresh_cache-df69e966-4580-42c5-8c88-53008807c1d0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.741802] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquired lock "refresh_cache-df69e966-4580-42c5-8c88-53008807c1d0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.742047] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 592.743411] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.444s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.743463] env[62521]: DEBUG nova.objects.instance [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62521) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 592.752221] env[62521]: DEBUG nova.compute.manager [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 592.794747] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 592.795376] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 592.795376] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 592.795501] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 592.795611] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 592.797821] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 592.799156] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 592.799156] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 592.799156] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 592.799325] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 592.799451] env[62521]: DEBUG nova.virt.hardware [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.801446] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec2fc296-bf52-48fc-86b3-d7f3c7bfe6cb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.810506] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102de592-fbf5-4a9a-99a1-ceff9bfbf7d9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.827604] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 592.833463] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Creating folder: Project (2f811c6ab74248ea9ae53722989fdd39). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 592.834473] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-729b4ff7-85fc-4c59-8825-65894a29ab6e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.846302] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Created folder: Project (2f811c6ab74248ea9ae53722989fdd39) in parent group-v282025. [ 592.846302] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Creating folder: Instances. Parent ref: group-v282039. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 592.846302] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c690228b-8687-4c8e-a9ec-d605dd47789f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.854025] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Created folder: Instances in parent group-v282039. [ 592.854325] env[62521]: DEBUG oslo.service.loopingcall [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.856049] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 592.856049] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67db479d-6bd1-4666-aba3-a091d41e60c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.871759] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 592.871759] env[62521]: value = "task-1318387" [ 592.871759] env[62521]: _type = "Task" [ 592.871759] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.881488] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318387, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.969164] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.281629] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.383716] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318387, 'name': CreateVM_Task, 'duration_secs': 0.299921} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.384065] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 593.384611] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.384937] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.385486] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 593.385862] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2ac0c35-7b5c-4675-9ba1-d67503676d25 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.392554] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 593.392554] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52887e71-0c23-541c-c8ba-56130b97968d" [ 593.392554] env[62521]: _type = "Task" [ 593.392554] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.400094] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52887e71-0c23-541c-c8ba-56130b97968d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.417772] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.764734] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b7240fe3-d4c6-49d3-8337-fd80d5539e2a tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.767376] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.550s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.901600] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52887e71-0c23-541c-c8ba-56130b97968d, 'name': SearchDatastore_Task, 'duration_secs': 0.027065} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.901928] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.902213] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 593.902825] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.902873] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.903052] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 593.903318] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69dd090a-81f8-4048-be5e-7544e9b74ec3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.911406] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 593.911581] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 593.912295] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17adc503-f73f-4467-ba21-caea72f8e46a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.920012] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 593.920012] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52db935a-e861-b126-7fe5-f023bd81c888" [ 593.920012] env[62521]: _type = "Task" [ 593.920012] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.924572] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Releasing lock "refresh_cache-df69e966-4580-42c5-8c88-53008807c1d0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.924773] env[62521]: DEBUG nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 593.924948] env[62521]: DEBUG nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.925127] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.926799] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52db935a-e861-b126-7fe5-f023bd81c888, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.964618] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.432686] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52db935a-e861-b126-7fe5-f023bd81c888, 'name': SearchDatastore_Task, 'duration_secs': 0.008537} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.432961] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbfdafd9-6776-46ba-8560-38a0ab39cb1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.438937] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 594.438937] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520ff5a8-bd96-3e3e-5def-142401c0ec07" [ 594.438937] env[62521]: _type = "Task" [ 594.438937] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.448371] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520ff5a8-bd96-3e3e-5def-142401c0ec07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.466938] env[62521]: DEBUG nova.network.neutron [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.685429] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39fb5e51-a7de-493e-ad60-15979b50e1af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.695067] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818c6e66-7f1b-4731-9c11-aa29d0f180cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.751018] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e5073f-3b6f-45cd-a8f1-6a114250286f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.757625] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8398246-5221-4dd6-a77e-64fd2f842584 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.772238] env[62521]: DEBUG nova.compute.provider_tree [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.948276] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520ff5a8-bd96-3e3e-5def-142401c0ec07, 'name': SearchDatastore_Task, 'duration_secs': 0.009064} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.948448] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.948704] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 20b01df0-4ced-4668-a786-046251a0c7ed/20b01df0-4ced-4668-a786-046251a0c7ed.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 594.949972] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-265bdc38-d58e-4de0-94df-b5fe59028f77 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.956100] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 594.956100] env[62521]: value = "task-1318388" [ 594.956100] env[62521]: _type = "Task" [ 594.956100] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.966166] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318388, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.975181] env[62521]: INFO nova.compute.manager [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: df69e966-4580-42c5-8c88-53008807c1d0] Took 1.05 seconds to deallocate network for instance. [ 595.278217] env[62521]: DEBUG nova.scheduler.client.report [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.467923] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318388, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.782172] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.014s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.782681] env[62521]: ERROR nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8842b0e6-a852-40b2-8c55-3dcb120fd6d3, please check neutron logs for more information. [ 595.782681] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] Traceback (most recent call last): [ 595.782681] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.782681] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] self.driver.spawn(context, instance, image_meta, [ 595.782681] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 595.782681] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.782681] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.782681] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] vm_ref = self.build_virtual_machine(instance, [ 595.782681] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.782681] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.782681] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] for vif in network_info: [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] return self._sync_wrapper(fn, *args, **kwargs) [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] self.wait() [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] self[:] = self._gt.wait() [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] return self._exit_event.wait() [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] result = hub.switch() [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.783451] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] return self.greenlet.switch() [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] result = function(*args, **kwargs) [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] return func(*args, **kwargs) [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] raise e [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] nwinfo = self.network_api.allocate_for_instance( [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] created_port_ids = self._update_ports_for_instance( [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] with excutils.save_and_reraise_exception(): [ 595.784129] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] self.force_reraise() [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] raise self.value [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] updated_port = self._update_port( [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] _ensure_no_port_binding_failure(port) [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] raise exception.PortBindingFailed(port_id=port['id']) [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] nova.exception.PortBindingFailed: Binding failed for port 8842b0e6-a852-40b2-8c55-3dcb120fd6d3, please check neutron logs for more information. [ 595.784599] env[62521]: ERROR nova.compute.manager [instance: da34c648-4074-494e-a569-e8e839207a70] [ 595.784991] env[62521]: DEBUG nova.compute.utils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Binding failed for port 8842b0e6-a852-40b2-8c55-3dcb120fd6d3, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 595.786109] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.345s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.787800] env[62521]: INFO nova.compute.claims [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.791953] env[62521]: DEBUG nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Build of instance da34c648-4074-494e-a569-e8e839207a70 was re-scheduled: Binding failed for port 8842b0e6-a852-40b2-8c55-3dcb120fd6d3, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 595.792429] env[62521]: DEBUG nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 595.793730] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Acquiring lock "refresh_cache-da34c648-4074-494e-a569-e8e839207a70" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.796016] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Acquired lock "refresh_cache-da34c648-4074-494e-a569-e8e839207a70" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.796016] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 595.973021] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318388, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590889} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.973021] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 20b01df0-4ced-4668-a786-046251a0c7ed/20b01df0-4ced-4668-a786-046251a0c7ed.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 595.973021] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 595.973021] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dedba08e-fd59-4474-be6a-f0f4dffbd79f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.978370] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 595.978370] env[62521]: value = "task-1318389" [ 595.978370] env[62521]: _type = "Task" [ 595.978370] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.988631] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318389, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.014414] env[62521]: INFO nova.scheduler.client.report [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Deleted allocations for instance df69e966-4580-42c5-8c88-53008807c1d0 [ 596.324649] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.484683] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.491176] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318389, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079402} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.492229] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 596.493185] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edc8a3d-6979-430d-be78-76f7fbaddbcf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.517308] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] 20b01df0-4ced-4668-a786-046251a0c7ed/20b01df0-4ced-4668-a786-046251a0c7ed.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 596.517959] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efc7a16c-b5d2-4670-88be-49586e6128ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.534585] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e868a778-4c11-4521-981f-86c6242503b4 tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "df69e966-4580-42c5-8c88-53008807c1d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.554s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.541506] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 596.541506] env[62521]: value = "task-1318390" [ 596.541506] env[62521]: _type = "Task" [ 596.541506] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.552595] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318390, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.988333] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Releasing lock "refresh_cache-da34c648-4074-494e-a569-e8e839207a70" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.988609] env[62521]: DEBUG nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 596.988765] env[62521]: DEBUG nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.988937] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 597.039277] env[62521]: DEBUG nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.052829] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318390, 'name': ReconfigVM_Task, 'duration_secs': 0.265892} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.053731] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Reconfigured VM instance instance-0000000c to attach disk [datastore1] 20b01df0-4ced-4668-a786-046251a0c7ed/20b01df0-4ced-4668-a786-046251a0c7ed.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 597.053867] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2dc6a62a-93fa-471f-a080-b2d1bc647fa6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.060959] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 597.060959] env[62521]: value = "task-1318391" [ 597.060959] env[62521]: _type = "Task" [ 597.060959] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.076831] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318391, 'name': Rename_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.108448] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.212371] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2dd1128-b33e-4085-9ae5-1b87babc54af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.220771] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2767c1-ecec-4d1d-899c-7479ba78d3ef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.264548] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce280333-a49a-46fc-9516-3ee7752a28b6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.273936] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba1dbf7-0473-48eb-a3b8-b6fee11a94c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.287316] env[62521]: DEBUG nova.compute.provider_tree [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.568766] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.574699] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318391, 'name': Rename_Task, 'duration_secs': 0.147072} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.574966] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 597.575595] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b876f20-80f9-48e2-8329-01a24400e500 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.582540] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 597.582540] env[62521]: value = "task-1318392" [ 597.582540] env[62521]: _type = "Task" [ 597.582540] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.592719] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.614456] env[62521]: DEBUG nova.network.neutron [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.791026] env[62521]: DEBUG nova.scheduler.client.report [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.879029] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Acquiring lock "a4965864-0127-4842-9478-76e369e2ecbe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.879029] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Lock "a4965864-0127-4842-9478-76e369e2ecbe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.094145] env[62521]: DEBUG oslo_vmware.api [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318392, 'name': PowerOnVM_Task, 'duration_secs': 0.441142} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.094507] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 598.094774] env[62521]: INFO nova.compute.manager [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Took 5.34 seconds to spawn the instance on the hypervisor. [ 598.095053] env[62521]: DEBUG nova.compute.manager [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 598.095977] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279189ea-9dcd-4725-bf48-616b71a2ceaf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.117712] env[62521]: INFO nova.compute.manager [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] [instance: da34c648-4074-494e-a569-e8e839207a70] Took 1.13 seconds to deallocate network for instance. [ 598.299911] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.300103] env[62521]: DEBUG nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 598.306503] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.308747] env[62521]: INFO nova.compute.claims [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 598.616308] env[62521]: INFO nova.compute.manager [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Took 32.86 seconds to build instance. [ 598.807759] env[62521]: DEBUG nova.compute.utils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.809155] env[62521]: DEBUG nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 598.809326] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 598.868802] env[62521]: DEBUG nova.policy [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '55fcd711563544e4b2dcf96b1765058c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42624a377e774a029e68f658517c8a70', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 599.118234] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dcf2b652-9f8c-411d-864c-10915e423270 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "20b01df0-4ced-4668-a786-046251a0c7ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.951s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.165692] env[62521]: INFO nova.scheduler.client.report [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Deleted allocations for instance da34c648-4074-494e-a569-e8e839207a70 [ 599.314576] env[62521]: DEBUG nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.377734] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Successfully created port: 5d67aeef-bad4-4299-bc25-6652e49a217f {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.621673] env[62521]: DEBUG nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 599.683020] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ed9f0ee-4fc0-4328-9a4d-d7caaf648615 tempest-VolumesAssistedSnapshotsTest-741690320 tempest-VolumesAssistedSnapshotsTest-741690320-project-member] Lock "da34c648-4074-494e-a569-e8e839207a70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.635s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.744445] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1443f6-d9bd-4ae1-bcd7-d4ee09810acb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.752494] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c7fbd6-ca45-4ec7-b463-d42a88f39e33 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.787870] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95631ef-bbd4-458d-93d9-779c10987d49 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.796560] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67ced41-e60a-41a5-9f9f-02acac9eeb29 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.812578] env[62521]: DEBUG nova.compute.provider_tree [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.856069] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquiring lock "cf71efc9-cfa1-4856-a733-5434bf4465b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.856069] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "cf71efc9-cfa1-4856-a733-5434bf4465b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.156879] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.186200] env[62521]: DEBUG nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 600.316397] env[62521]: DEBUG nova.scheduler.client.report [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.333764] env[62521]: DEBUG nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.367443] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:30:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1612493991',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-877688907',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.367696] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.369037] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.369037] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.369037] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.369037] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.369037] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.369279] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.369279] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.369279] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.369279] env[62521]: DEBUG nova.virt.hardware [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.370163] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ef9b66-8994-472a-9bfe-8f9f9ba30173 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.379527] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4bbfdd-e2ea-4e3a-83c9-82a9ba354dbd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.618206] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Acquiring lock "d7575b46-ecf9-4154-92b1-86119e727dda" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.618206] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Lock "d7575b46-ecf9-4154-92b1-86119e727dda" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.663186] env[62521]: ERROR nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5d67aeef-bad4-4299-bc25-6652e49a217f, please check neutron logs for more information. [ 600.663186] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 600.663186] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.663186] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 600.663186] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.663186] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 600.663186] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.663186] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 600.663186] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.663186] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 600.663186] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.663186] env[62521]: ERROR nova.compute.manager raise self.value [ 600.663186] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.663186] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 600.663186] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.663186] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 600.663782] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.663782] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 600.663782] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5d67aeef-bad4-4299-bc25-6652e49a217f, please check neutron logs for more information. [ 600.663782] env[62521]: ERROR nova.compute.manager [ 600.663782] env[62521]: Traceback (most recent call last): [ 600.663782] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 600.663782] env[62521]: listener.cb(fileno) [ 600.663782] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.663782] env[62521]: result = function(*args, **kwargs) [ 600.663782] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 600.663782] env[62521]: return func(*args, **kwargs) [ 600.663782] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.663782] env[62521]: raise e [ 600.663782] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.663782] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 600.663782] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.663782] env[62521]: created_port_ids = self._update_ports_for_instance( [ 600.663782] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.663782] env[62521]: with excutils.save_and_reraise_exception(): [ 600.663782] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.663782] env[62521]: self.force_reraise() [ 600.663782] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.663782] env[62521]: raise self.value [ 600.663782] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.663782] env[62521]: updated_port = self._update_port( [ 600.663782] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.663782] env[62521]: _ensure_no_port_binding_failure(port) [ 600.663782] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.663782] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 600.665193] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 5d67aeef-bad4-4299-bc25-6652e49a217f, please check neutron logs for more information. [ 600.665193] env[62521]: Removing descriptor: 18 [ 600.665193] env[62521]: ERROR nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5d67aeef-bad4-4299-bc25-6652e49a217f, please check neutron logs for more information. [ 600.665193] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Traceback (most recent call last): [ 600.665193] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 600.665193] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] yield resources [ 600.665193] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.665193] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] self.driver.spawn(context, instance, image_meta, [ 600.665193] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 600.665193] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.665193] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.665193] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] vm_ref = self.build_virtual_machine(instance, [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] for vif in network_info: [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] return self._sync_wrapper(fn, *args, **kwargs) [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] self.wait() [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] self[:] = self._gt.wait() [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] return self._exit_event.wait() [ 600.666021] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] result = hub.switch() [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] return self.greenlet.switch() [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] result = function(*args, **kwargs) [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] return func(*args, **kwargs) [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] raise e [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] nwinfo = self.network_api.allocate_for_instance( [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.666437] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] created_port_ids = self._update_ports_for_instance( [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] with excutils.save_and_reraise_exception(): [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] self.force_reraise() [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] raise self.value [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] updated_port = self._update_port( [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] _ensure_no_port_binding_failure(port) [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.667411] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] raise exception.PortBindingFailed(port_id=port['id']) [ 600.667782] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] nova.exception.PortBindingFailed: Binding failed for port 5d67aeef-bad4-4299-bc25-6652e49a217f, please check neutron logs for more information. [ 600.667782] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] [ 600.667782] env[62521]: INFO nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Terminating instance [ 600.669010] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Acquiring lock "refresh_cache-6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.669010] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Acquired lock "refresh_cache-6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.669010] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.718667] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.821499] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.824419] env[62521]: DEBUG nova.compute.manager [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.825902] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.245s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.827141] env[62521]: INFO nova.compute.claims [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 601.193931] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.218189] env[62521]: DEBUG nova.compute.manager [req-7ed7ac23-99c8-492b-83ae-cbcf5f4dc9f0 req-3a9dc3e0-0ade-4a4f-97d6-02586159b478 service nova] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Received event network-changed-5d67aeef-bad4-4299-bc25-6652e49a217f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.218189] env[62521]: DEBUG nova.compute.manager [req-7ed7ac23-99c8-492b-83ae-cbcf5f4dc9f0 req-3a9dc3e0-0ade-4a4f-97d6-02586159b478 service nova] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Refreshing instance network info cache due to event network-changed-5d67aeef-bad4-4299-bc25-6652e49a217f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 601.218189] env[62521]: DEBUG oslo_concurrency.lockutils [req-7ed7ac23-99c8-492b-83ae-cbcf5f4dc9f0 req-3a9dc3e0-0ade-4a4f-97d6-02586159b478 service nova] Acquiring lock "refresh_cache-6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.306146] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.336891] env[62521]: DEBUG nova.compute.utils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 601.338741] env[62521]: DEBUG nova.compute.manager [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Not allocating networking since 'none' was specified. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 601.552934] env[62521]: INFO nova.compute.manager [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Rebuilding instance [ 601.607358] env[62521]: DEBUG nova.compute.manager [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 601.611033] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192664a6-a5f5-41bc-a5bf-cf85c72a2f2c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.713511] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "0d72d988-f96b-428c-89bf-a96572f71db5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.713748] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "0d72d988-f96b-428c-89bf-a96572f71db5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.809842] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Releasing lock "refresh_cache-6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.809842] env[62521]: DEBUG nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 601.809842] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 601.809842] env[62521]: DEBUG oslo_concurrency.lockutils [req-7ed7ac23-99c8-492b-83ae-cbcf5f4dc9f0 req-3a9dc3e0-0ade-4a4f-97d6-02586159b478 service nova] Acquired lock "refresh_cache-6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.809842] env[62521]: DEBUG nova.network.neutron [req-7ed7ac23-99c8-492b-83ae-cbcf5f4dc9f0 req-3a9dc3e0-0ade-4a4f-97d6-02586159b478 service nova] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Refreshing network info cache for port 5d67aeef-bad4-4299-bc25-6652e49a217f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 601.811720] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ec13cb4-48db-4c47-bc9e-811d1936facb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.820958] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa00a8ad-02b3-480e-abe7-b451fd92d24c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.846132] env[62521]: DEBUG nova.compute.manager [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.849841] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4 could not be found. [ 601.850057] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 601.850383] env[62521]: INFO nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 601.850654] env[62521]: DEBUG oslo.service.loopingcall [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.852615] env[62521]: DEBUG nova.compute.manager [-] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.852615] env[62521]: DEBUG nova.network.neutron [-] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.868514] env[62521]: DEBUG nova.network.neutron [-] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.123899] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 602.124674] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-669fcddf-6ec3-4f2b-8dfe-e4d3a13bb331 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.136133] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 602.136133] env[62521]: value = "task-1318393" [ 602.136133] env[62521]: _type = "Task" [ 602.136133] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.153463] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.281598] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d0bb6b-0dec-4465-8dcf-584bce3d37bb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.290268] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9f8b87-6f50-4764-b92b-1a7bd43edbc4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.323692] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6955cbda-2fb5-482f-9b0d-cc73c9fe3556 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.335091] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30de353-0f56-4fa7-80fe-880c47ad3073 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.351237] env[62521]: DEBUG nova.compute.provider_tree [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.353382] env[62521]: DEBUG nova.network.neutron [req-7ed7ac23-99c8-492b-83ae-cbcf5f4dc9f0 req-3a9dc3e0-0ade-4a4f-97d6-02586159b478 service nova] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.370774] env[62521]: DEBUG nova.network.neutron [-] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.465212] env[62521]: DEBUG nova.network.neutron [req-7ed7ac23-99c8-492b-83ae-cbcf5f4dc9f0 req-3a9dc3e0-0ade-4a4f-97d6-02586159b478 service nova] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.582213] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "884d2f4f-c0a2-4d08-962e-d6347dc9f678" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.582828] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "884d2f4f-c0a2-4d08-962e-d6347dc9f678" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.585740] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "5fa1c73e-3074-4fe0-8b4d-926bf02f1453" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.585956] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "5fa1c73e-3074-4fe0-8b4d-926bf02f1453" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.618625] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "79b55308-ff33-4b2a-bca8-30a1c5cf5916" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.618857] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "79b55308-ff33-4b2a-bca8-30a1c5cf5916" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.648821] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318393, 'name': PowerOffVM_Task, 'duration_secs': 0.123431} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.649537] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 602.649837] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.652171] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adae4d5c-7240-4d38-8c69-1ec488d7699e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.659549] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "5adbe047-7055-428f-b72b-67023c68cc04" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.659549] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "5adbe047-7055-428f-b72b-67023c68cc04" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.667179] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 602.667179] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb9f911b-53c6-4511-9a38-d6543f934379 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.693901] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 602.693901] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 602.693901] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Deleting the datastore file [datastore1] 20b01df0-4ced-4668-a786-046251a0c7ed {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 602.693901] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3757b196-198b-49a9-8a89-747f1038de64 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.699904] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 602.699904] env[62521]: value = "task-1318395" [ 602.699904] env[62521]: _type = "Task" [ 602.699904] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.711539] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318395, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.950674] env[62521]: DEBUG nova.scheduler.client.report [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.950674] env[62521]: DEBUG nova.compute.manager [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.950674] env[62521]: INFO nova.compute.manager [-] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Took 1.02 seconds to deallocate network for instance. [ 602.950674] env[62521]: DEBUG nova.compute.claims [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 602.952224] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.952224] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.952224] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.952745] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.952745] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.952745] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.952745] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.952745] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.953624] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.953624] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.953624] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.953624] env[62521]: DEBUG nova.virt.hardware [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.953624] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33cab432-5c4c-439d-9074-0988451b1abb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.953624] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de41964-12e0-4731-acd5-0f3ef0adb345 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.958267] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 602.958267] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Creating folder: Project (c80cf3f2edc34258bc674d696d9e7ec0). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 602.958267] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f8dd857-a759-4030-81f4-9d4ce13f6d5d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.958267] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Created folder: Project (c80cf3f2edc34258bc674d696d9e7ec0) in parent group-v282025. [ 602.958267] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Creating folder: Instances. Parent ref: group-v282042. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 602.958267] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d29b1f61-d399-4eb2-b669-9e4ebef91da8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.959430] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Created folder: Instances in parent group-v282042. [ 602.959430] env[62521]: DEBUG oslo.service.loopingcall [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.959430] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 602.959430] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9701e4d-d67e-423d-862d-d4b673c9ee21 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.959430] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 602.959430] env[62521]: value = "task-1318398" [ 602.959430] env[62521]: _type = "Task" [ 602.959430] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.965104] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318398, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.969056] env[62521]: DEBUG oslo_concurrency.lockutils [req-7ed7ac23-99c8-492b-83ae-cbcf5f4dc9f0 req-3a9dc3e0-0ade-4a4f-97d6-02586159b478 service nova] Releasing lock "refresh_cache-6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.212318] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318395, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.115731} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.212608] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 603.212786] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 603.212959] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.313742] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "72e3b843-c54a-4de7-ace9-ee25c5df14bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.317012] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "72e3b843-c54a-4de7-ace9-ee25c5df14bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.367570] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.368061] env[62521]: DEBUG nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 603.370534] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.372019] env[62521]: INFO nova.compute.claims [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 603.467031] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318398, 'name': CreateVM_Task, 'duration_secs': 0.285312} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.467162] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 603.467474] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.467635] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.467966] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 603.468221] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-141527b6-2545-4ecd-9b0d-c03d1f275eef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.472522] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for the task: (returnval){ [ 603.472522] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ae0f1b-f6c4-e55b-d977-c243262305d1" [ 603.472522] env[62521]: _type = "Task" [ 603.472522] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.483755] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ae0f1b-f6c4-e55b-d977-c243262305d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.519240] env[62521]: DEBUG nova.compute.manager [req-af33ef68-49a5-401e-97a1-71798c63e1f2 req-f479e89f-0ce2-4d2c-b122-d4adffb236df service nova] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Received event network-vif-deleted-5d67aeef-bad4-4299-bc25-6652e49a217f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.876153] env[62521]: DEBUG nova.compute.utils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 603.880741] env[62521]: DEBUG nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 603.880842] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 603.947313] env[62521]: DEBUG nova.policy [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd1fd76f1323e48f88490eac8cc96b2eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd670164b9d524f15911efc1681ac4bce', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 603.983615] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ae0f1b-f6c4-e55b-d977-c243262305d1, 'name': SearchDatastore_Task, 'duration_secs': 0.009606} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.984134] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.984491] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 603.984843] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.985129] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.985432] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 603.985791] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c68a7fc9-04e5-492c-bf96-aaf07061baa9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.994901] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 603.995272] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 603.996139] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3beb8cea-440a-4d5c-bf7a-ccbfc07bf6b8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.002762] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for the task: (returnval){ [ 604.002762] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52639a76-2193-9bdd-7603-68f3fbb45039" [ 604.002762] env[62521]: _type = "Task" [ 604.002762] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.011570] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52639a76-2193-9bdd-7603-68f3fbb45039, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.249240] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.249240] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.249240] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.249240] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.249442] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.249807] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.250172] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.250455] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.250854] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.251843] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.252166] env[62521]: DEBUG nova.virt.hardware [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.253163] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf56ddd-1895-41d9-8094-6bfdf542c81d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.261941] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bc4bb0-54a3-4a8f-aa17-eb378e572c15 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.276045] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 604.283219] env[62521]: DEBUG oslo.service.loopingcall [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.283542] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 604.284143] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5bea03f4-ae88-4f5a-90f8-d4941e36ffe9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.302170] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 604.302170] env[62521]: value = "task-1318399" [ 604.302170] env[62521]: _type = "Task" [ 604.302170] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.302597] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Successfully created port: 5e484068-db21-4dc6-9171-8e256d0e84fa {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.312140] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318399, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.384013] env[62521]: DEBUG nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.519488] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52639a76-2193-9bdd-7603-68f3fbb45039, 'name': SearchDatastore_Task, 'duration_secs': 0.00873} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.520278] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e33f63b2-7b14-4eab-87ba-7305e2a35980 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.531471] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for the task: (returnval){ [ 604.531471] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a38fff-15a4-17db-eb46-2dcdc623f9da" [ 604.531471] env[62521]: _type = "Task" [ 604.531471] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.539881] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a38fff-15a4-17db-eb46-2dcdc623f9da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.819048] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318399, 'name': CreateVM_Task, 'duration_secs': 0.311733} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.819368] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 604.819700] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.820358] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.820358] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 604.820732] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e347b0f2-b144-40c0-8ce1-25775785a795 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.827020] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 604.827020] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ce5110-cff7-1fed-a7da-46d800429aae" [ 604.827020] env[62521]: _type = "Task" [ 604.827020] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.835401] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ce5110-cff7-1fed-a7da-46d800429aae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.855035] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f7be0b-f994-4cf6-a4f9-ac308b33cea5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.862057] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7de82f8-a485-415a-b533-27d0c9f9e23f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.898355] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7262413-da9a-40a5-b2bc-4762af1549bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.906407] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db62098-bb63-4463-b00c-7af115259407 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.920087] env[62521]: DEBUG nova.compute.provider_tree [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.041451] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a38fff-15a4-17db-eb46-2dcdc623f9da, 'name': SearchDatastore_Task, 'duration_secs': 0.008824} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.041842] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.042019] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 7653f469-5ce5-4d74-9a3f-55a6adb3a255/7653f469-5ce5-4d74-9a3f-55a6adb3a255.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 605.042323] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f37b5490-6ce0-4982-835a-7f86510c9d4a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.052125] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for the task: (returnval){ [ 605.052125] env[62521]: value = "task-1318400" [ 605.052125] env[62521]: _type = "Task" [ 605.052125] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.059506] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318400, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.300277] env[62521]: ERROR nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5e484068-db21-4dc6-9171-8e256d0e84fa, please check neutron logs for more information. [ 605.300277] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.300277] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.300277] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.300277] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.300277] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.300277] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.300277] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.300277] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.300277] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 605.300277] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.300277] env[62521]: ERROR nova.compute.manager raise self.value [ 605.300277] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.300277] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.300277] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.300277] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.300960] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.300960] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.300960] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5e484068-db21-4dc6-9171-8e256d0e84fa, please check neutron logs for more information. [ 605.300960] env[62521]: ERROR nova.compute.manager [ 605.300960] env[62521]: Traceback (most recent call last): [ 605.300960] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.300960] env[62521]: listener.cb(fileno) [ 605.300960] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.300960] env[62521]: result = function(*args, **kwargs) [ 605.300960] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 605.300960] env[62521]: return func(*args, **kwargs) [ 605.300960] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.300960] env[62521]: raise e [ 605.300960] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.300960] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 605.300960] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.300960] env[62521]: created_port_ids = self._update_ports_for_instance( [ 605.300960] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.300960] env[62521]: with excutils.save_and_reraise_exception(): [ 605.300960] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.300960] env[62521]: self.force_reraise() [ 605.300960] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.300960] env[62521]: raise self.value [ 605.300960] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.300960] env[62521]: updated_port = self._update_port( [ 605.300960] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.300960] env[62521]: _ensure_no_port_binding_failure(port) [ 605.300960] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.300960] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.301931] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 5e484068-db21-4dc6-9171-8e256d0e84fa, please check neutron logs for more information. [ 605.301931] env[62521]: Removing descriptor: 18 [ 605.337332] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ce5110-cff7-1fed-a7da-46d800429aae, 'name': SearchDatastore_Task, 'duration_secs': 0.010682} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.337888] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.338238] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 605.338539] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.338739] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.338971] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 605.339435] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb381973-0f00-4858-aabe-0e4a2a558b0d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.354487] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 605.354682] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 605.355468] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdc2e1ea-e3d6-4df7-8217-df512bbf7797 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.361920] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 605.361920] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52506661-48ff-dcd0-fe50-499107bb41ed" [ 605.361920] env[62521]: _type = "Task" [ 605.361920] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.371803] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52506661-48ff-dcd0-fe50-499107bb41ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.403270] env[62521]: DEBUG nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 605.425813] env[62521]: DEBUG nova.scheduler.client.report [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.443417] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 605.443417] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 605.443417] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 605.443862] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 605.443862] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 605.443862] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 605.443862] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 605.443862] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 605.444052] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 605.444052] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 605.444052] env[62521]: DEBUG nova.virt.hardware [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.444508] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6726185b-354e-4af8-b59d-4b1334f99d7f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.453539] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99663bb7-0cec-4e3c-9e09-30ec5f341e64 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.472018] env[62521]: ERROR nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5e484068-db21-4dc6-9171-8e256d0e84fa, please check neutron logs for more information. [ 605.472018] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Traceback (most recent call last): [ 605.472018] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 605.472018] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] yield resources [ 605.472018] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.472018] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] self.driver.spawn(context, instance, image_meta, [ 605.472018] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 605.472018] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.472018] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.472018] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] vm_ref = self.build_virtual_machine(instance, [ 605.472018] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] for vif in network_info: [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] return self._sync_wrapper(fn, *args, **kwargs) [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] self.wait() [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] self[:] = self._gt.wait() [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] return self._exit_event.wait() [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 605.472454] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] current.throw(*self._exc) [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] result = function(*args, **kwargs) [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] return func(*args, **kwargs) [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] raise e [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] nwinfo = self.network_api.allocate_for_instance( [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] created_port_ids = self._update_ports_for_instance( [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] with excutils.save_and_reraise_exception(): [ 605.472883] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] self.force_reraise() [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] raise self.value [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] updated_port = self._update_port( [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] _ensure_no_port_binding_failure(port) [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] raise exception.PortBindingFailed(port_id=port['id']) [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] nova.exception.PortBindingFailed: Binding failed for port 5e484068-db21-4dc6-9171-8e256d0e84fa, please check neutron logs for more information. [ 605.473341] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] [ 605.473341] env[62521]: INFO nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Terminating instance [ 605.477050] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquiring lock "refresh_cache-6296d4ea-d678-4981-bc5a-729e83020fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.477050] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquired lock "refresh_cache-6296d4ea-d678-4981-bc5a-729e83020fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.477050] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.562468] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318400, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.724493] env[62521]: DEBUG nova.compute.manager [req-ce16278a-05e4-46b9-ae13-be569a33ff62 req-5bab8657-b700-4e0b-a53e-6fa587fb1e68 service nova] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Received event network-changed-5e484068-db21-4dc6-9171-8e256d0e84fa {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.724493] env[62521]: DEBUG nova.compute.manager [req-ce16278a-05e4-46b9-ae13-be569a33ff62 req-5bab8657-b700-4e0b-a53e-6fa587fb1e68 service nova] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Refreshing instance network info cache due to event network-changed-5e484068-db21-4dc6-9171-8e256d0e84fa. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 605.724493] env[62521]: DEBUG oslo_concurrency.lockutils [req-ce16278a-05e4-46b9-ae13-be569a33ff62 req-5bab8657-b700-4e0b-a53e-6fa587fb1e68 service nova] Acquiring lock "refresh_cache-6296d4ea-d678-4981-bc5a-729e83020fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.824030] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquiring lock "86727bdc-342a-429c-85ce-adec8823b799" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.824030] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "86727bdc-342a-429c-85ce-adec8823b799" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.873081] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52506661-48ff-dcd0-fe50-499107bb41ed, 'name': SearchDatastore_Task, 'duration_secs': 0.08173} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.873950] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a050fd3d-14a6-4c1b-9ebb-876d6ec95226 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.879633] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 605.879633] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529d0dd3-324a-a4ed-f21d-7eecb595d463" [ 605.879633] env[62521]: _type = "Task" [ 605.879633] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.887499] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529d0dd3-324a-a4ed-f21d-7eecb595d463, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.932066] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.934320] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 605.941884] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.573s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.943401] env[62521]: INFO nova.compute.claims [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.006334] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.061638] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318400, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.61817} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.061901] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 7653f469-5ce5-4d74-9a3f-55a6adb3a255/7653f469-5ce5-4d74-9a3f-55a6adb3a255.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 606.062119] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 606.062364] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd4d757c-8cf1-45de-9d33-36d50a6ff55e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.068699] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for the task: (returnval){ [ 606.068699] env[62521]: value = "task-1318401" [ 606.068699] env[62521]: _type = "Task" [ 606.068699] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.076223] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318401, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.088074] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.393889] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529d0dd3-324a-a4ed-f21d-7eecb595d463, 'name': SearchDatastore_Task, 'duration_secs': 0.009405} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.394403] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.394807] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 20b01df0-4ced-4668-a786-046251a0c7ed/20b01df0-4ced-4668-a786-046251a0c7ed.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 606.396040] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d29527a1-37b8-4c0a-a25a-f7aa94f37112 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.402646] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 606.402646] env[62521]: value = "task-1318402" [ 606.402646] env[62521]: _type = "Task" [ 606.402646] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.410830] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318402, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.449367] env[62521]: DEBUG nova.compute.utils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 606.453154] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 606.453325] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 606.520392] env[62521]: DEBUG nova.policy [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05e0e43e72b74039b813d71aebd21872', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63ec3a6586ee49059b51ccfcaaeab7b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 606.578301] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318401, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069236} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.578872] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 606.579741] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29e491d-66e9-4c3a-9a8a-6bf6a86098d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.602228] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Reconfiguring VM instance instance-0000000e to attach disk [datastore1] 7653f469-5ce5-4d74-9a3f-55a6adb3a255/7653f469-5ce5-4d74-9a3f-55a6adb3a255.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 606.603091] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Releasing lock "refresh_cache-6296d4ea-d678-4981-bc5a-729e83020fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.604014] env[62521]: DEBUG nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 606.604398] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 606.605118] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b531f11-6517-45dc-8f1d-b598e3aff32c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.624481] env[62521]: DEBUG oslo_concurrency.lockutils [req-ce16278a-05e4-46b9-ae13-be569a33ff62 req-5bab8657-b700-4e0b-a53e-6fa587fb1e68 service nova] Acquired lock "refresh_cache-6296d4ea-d678-4981-bc5a-729e83020fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.624481] env[62521]: DEBUG nova.network.neutron [req-ce16278a-05e4-46b9-ae13-be569a33ff62 req-5bab8657-b700-4e0b-a53e-6fa587fb1e68 service nova] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Refreshing network info cache for port 5e484068-db21-4dc6-9171-8e256d0e84fa {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 606.625128] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4ff2f1bc-60d0-4567-aa8f-16dadc3e022e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.635232] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2fbb78-1528-4687-a7a6-2a053397b93e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.647164] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for the task: (returnval){ [ 606.647164] env[62521]: value = "task-1318403" [ 606.647164] env[62521]: _type = "Task" [ 606.647164] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.656917] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.662360] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6296d4ea-d678-4981-bc5a-729e83020fa2 could not be found. [ 606.662583] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 606.662760] env[62521]: INFO nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Took 0.06 seconds to destroy the instance on the hypervisor. [ 606.663012] env[62521]: DEBUG oslo.service.loopingcall [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.663235] env[62521]: DEBUG nova.compute.manager [-] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.663323] env[62521]: DEBUG nova.network.neutron [-] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.689524] env[62521]: DEBUG nova.network.neutron [-] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.914690] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318402, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.954296] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.044837] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Successfully created port: 7b451f88-f19d-4f24-9f30-a78410e4db4a {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.157570] env[62521]: DEBUG nova.network.neutron [req-ce16278a-05e4-46b9-ae13-be569a33ff62 req-5bab8657-b700-4e0b-a53e-6fa587fb1e68 service nova] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.168782] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318403, 'name': ReconfigVM_Task, 'duration_secs': 0.337498} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.168782] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Reconfigured VM instance instance-0000000e to attach disk [datastore1] 7653f469-5ce5-4d74-9a3f-55a6adb3a255/7653f469-5ce5-4d74-9a3f-55a6adb3a255.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 607.169220] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d59e7dcc-af1d-4cf6-9236-7e09ff438028 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.182386] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for the task: (returnval){ [ 607.182386] env[62521]: value = "task-1318404" [ 607.182386] env[62521]: _type = "Task" [ 607.182386] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.193432] env[62521]: DEBUG nova.network.neutron [-] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.201445] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318404, 'name': Rename_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.236137] env[62521]: DEBUG nova.network.neutron [req-ce16278a-05e4-46b9-ae13-be569a33ff62 req-5bab8657-b700-4e0b-a53e-6fa587fb1e68 service nova] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.415567] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318402, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572871} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.415854] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 20b01df0-4ced-4668-a786-046251a0c7ed/20b01df0-4ced-4668-a786-046251a0c7ed.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 607.416039] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 607.416279] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fea72c07-b398-4988-b1e3-730242ee512c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.429140] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 607.429140] env[62521]: value = "task-1318405" [ 607.429140] env[62521]: _type = "Task" [ 607.429140] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.438726] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318405, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.502785] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe319e5-7be2-4b0f-9c8a-8cfe470440f2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.512288] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a74ebfb-5d80-4854-a741-0847219a02ef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.543683] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ea0bd1-0fc6-4067-b0cb-cd684f2c2211 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.550987] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3f6310-57c2-441b-a9c7-14c93b6c4415 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.563990] env[62521]: DEBUG nova.compute.provider_tree [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.697743] env[62521]: INFO nova.compute.manager [-] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Took 1.03 seconds to deallocate network for instance. [ 607.698083] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318404, 'name': Rename_Task, 'duration_secs': 0.13516} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.700271] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 607.700271] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e76176eb-4a2a-4701-ba6b-3ed23f9161f6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.701969] env[62521]: DEBUG nova.compute.claims [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 607.702157] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.707072] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for the task: (returnval){ [ 607.707072] env[62521]: value = "task-1318406" [ 607.707072] env[62521]: _type = "Task" [ 607.707072] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.714800] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318406, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.737737] env[62521]: DEBUG oslo_concurrency.lockutils [req-ce16278a-05e4-46b9-ae13-be569a33ff62 req-5bab8657-b700-4e0b-a53e-6fa587fb1e68 service nova] Releasing lock "refresh_cache-6296d4ea-d678-4981-bc5a-729e83020fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.738125] env[62521]: DEBUG nova.compute.manager [req-ce16278a-05e4-46b9-ae13-be569a33ff62 req-5bab8657-b700-4e0b-a53e-6fa587fb1e68 service nova] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Received event network-vif-deleted-5e484068-db21-4dc6-9171-8e256d0e84fa {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 607.939689] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318405, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066267} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.940270] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 607.941127] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603f5ba2-d012-4320-b776-6f313717e779 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.969576] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] 20b01df0-4ced-4668-a786-046251a0c7ed/20b01df0-4ced-4668-a786-046251a0c7ed.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 607.969928] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e473feb7-bb76-4462-a0bd-08cb73ace887 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.987466] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 607.995547] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 607.995547] env[62521]: value = "task-1318407" [ 607.995547] env[62521]: _type = "Task" [ 607.995547] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.004957] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318407, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.026876] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.027272] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.027351] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.027519] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.027882] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.028186] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.028403] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.028638] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.029009] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.029280] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.029464] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.030553] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626d43d2-6b11-40aa-9ac1-5ea9cc60b464 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.040157] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f13bf1-9ff1-4217-b656-e59ef89ac09a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.067582] env[62521]: DEBUG nova.scheduler.client.report [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.097452] env[62521]: DEBUG nova.compute.manager [req-7cdb1335-b5fa-44f6-8d6d-2c544d5f0404 req-1dd24d4c-d933-481c-a008-7001f4d5aed9 service nova] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Received event network-changed-7b451f88-f19d-4f24-9f30-a78410e4db4a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.097972] env[62521]: DEBUG nova.compute.manager [req-7cdb1335-b5fa-44f6-8d6d-2c544d5f0404 req-1dd24d4c-d933-481c-a008-7001f4d5aed9 service nova] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Refreshing instance network info cache due to event network-changed-7b451f88-f19d-4f24-9f30-a78410e4db4a. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 608.097972] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cdb1335-b5fa-44f6-8d6d-2c544d5f0404 req-1dd24d4c-d933-481c-a008-7001f4d5aed9 service nova] Acquiring lock "refresh_cache-4b2fda93-a60d-4050-ae05-7a37ac284917" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.098310] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cdb1335-b5fa-44f6-8d6d-2c544d5f0404 req-1dd24d4c-d933-481c-a008-7001f4d5aed9 service nova] Acquired lock "refresh_cache-4b2fda93-a60d-4050-ae05-7a37ac284917" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.098310] env[62521]: DEBUG nova.network.neutron [req-7cdb1335-b5fa-44f6-8d6d-2c544d5f0404 req-1dd24d4c-d933-481c-a008-7001f4d5aed9 service nova] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Refreshing network info cache for port 7b451f88-f19d-4f24-9f30-a78410e4db4a {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 608.224031] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318406, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.235861] env[62521]: ERROR nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7b451f88-f19d-4f24-9f30-a78410e4db4a, please check neutron logs for more information. [ 608.235861] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 608.235861] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.235861] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 608.235861] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.235861] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 608.235861] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.235861] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 608.235861] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.235861] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 608.235861] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.235861] env[62521]: ERROR nova.compute.manager raise self.value [ 608.235861] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.235861] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 608.235861] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.235861] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 608.236741] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.236741] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 608.236741] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7b451f88-f19d-4f24-9f30-a78410e4db4a, please check neutron logs for more information. [ 608.236741] env[62521]: ERROR nova.compute.manager [ 608.236741] env[62521]: Traceback (most recent call last): [ 608.236741] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 608.236741] env[62521]: listener.cb(fileno) [ 608.236741] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.236741] env[62521]: result = function(*args, **kwargs) [ 608.236741] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 608.236741] env[62521]: return func(*args, **kwargs) [ 608.236741] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.236741] env[62521]: raise e [ 608.236741] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.236741] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 608.236741] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.236741] env[62521]: created_port_ids = self._update_ports_for_instance( [ 608.236741] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.236741] env[62521]: with excutils.save_and_reraise_exception(): [ 608.236741] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.236741] env[62521]: self.force_reraise() [ 608.236741] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.236741] env[62521]: raise self.value [ 608.236741] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.236741] env[62521]: updated_port = self._update_port( [ 608.236741] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.236741] env[62521]: _ensure_no_port_binding_failure(port) [ 608.236741] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.236741] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 608.238227] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 7b451f88-f19d-4f24-9f30-a78410e4db4a, please check neutron logs for more information. [ 608.238227] env[62521]: Removing descriptor: 18 [ 608.238227] env[62521]: ERROR nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7b451f88-f19d-4f24-9f30-a78410e4db4a, please check neutron logs for more information. [ 608.238227] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Traceback (most recent call last): [ 608.238227] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 608.238227] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] yield resources [ 608.238227] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.238227] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] self.driver.spawn(context, instance, image_meta, [ 608.238227] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 608.238227] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.238227] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.238227] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] vm_ref = self.build_virtual_machine(instance, [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] for vif in network_info: [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] return self._sync_wrapper(fn, *args, **kwargs) [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] self.wait() [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] self[:] = self._gt.wait() [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] return self._exit_event.wait() [ 608.238744] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] result = hub.switch() [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] return self.greenlet.switch() [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] result = function(*args, **kwargs) [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] return func(*args, **kwargs) [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] raise e [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] nwinfo = self.network_api.allocate_for_instance( [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.239234] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] created_port_ids = self._update_ports_for_instance( [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] with excutils.save_and_reraise_exception(): [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] self.force_reraise() [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] raise self.value [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] updated_port = self._update_port( [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] _ensure_no_port_binding_failure(port) [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.239793] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] raise exception.PortBindingFailed(port_id=port['id']) [ 608.240629] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] nova.exception.PortBindingFailed: Binding failed for port 7b451f88-f19d-4f24-9f30-a78410e4db4a, please check neutron logs for more information. [ 608.240629] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] [ 608.240629] env[62521]: INFO nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Terminating instance [ 608.240629] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "refresh_cache-4b2fda93-a60d-4050-ae05-7a37ac284917" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.505929] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318407, 'name': ReconfigVM_Task, 'duration_secs': 0.292323} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.506395] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Reconfigured VM instance instance-0000000c to attach disk [datastore1] 20b01df0-4ced-4668-a786-046251a0c7ed/20b01df0-4ced-4668-a786-046251a0c7ed.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 608.506996] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ee58e88-cb5d-4e03-955b-fd3b420ddf50 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.513314] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 608.513314] env[62521]: value = "task-1318408" [ 608.513314] env[62521]: _type = "Task" [ 608.513314] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.522972] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318408, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.573043] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.631s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.573657] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.576543] env[62521]: DEBUG oslo_concurrency.lockutils [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.833s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.576708] env[62521]: DEBUG nova.objects.instance [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62521) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 608.624952] env[62521]: DEBUG nova.network.neutron [req-7cdb1335-b5fa-44f6-8d6d-2c544d5f0404 req-1dd24d4c-d933-481c-a008-7001f4d5aed9 service nova] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.708168] env[62521]: DEBUG nova.network.neutron [req-7cdb1335-b5fa-44f6-8d6d-2c544d5f0404 req-1dd24d4c-d933-481c-a008-7001f4d5aed9 service nova] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.719859] env[62521]: DEBUG oslo_vmware.api [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318406, 'name': PowerOnVM_Task, 'duration_secs': 0.540657} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.721230] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 608.721536] env[62521]: INFO nova.compute.manager [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Took 5.86 seconds to spawn the instance on the hypervisor. [ 608.721759] env[62521]: DEBUG nova.compute.manager [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 608.722566] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce758b35-4a02-448d-9069-15ff20c3a861 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.023751] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318408, 'name': Rename_Task, 'duration_secs': 0.130187} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.026247] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 609.026247] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6c95b30-9e2d-44b2-966f-4ef0551ef2bf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.033659] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 609.033659] env[62521]: value = "task-1318409" [ 609.033659] env[62521]: _type = "Task" [ 609.033659] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.043123] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318409, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.084024] env[62521]: DEBUG nova.compute.utils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.086544] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.086981] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 609.162968] env[62521]: DEBUG nova.policy [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05e0e43e72b74039b813d71aebd21872', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63ec3a6586ee49059b51ccfcaaeab7b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 609.215509] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cdb1335-b5fa-44f6-8d6d-2c544d5f0404 req-1dd24d4c-d933-481c-a008-7001f4d5aed9 service nova] Releasing lock "refresh_cache-4b2fda93-a60d-4050-ae05-7a37ac284917" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.217455] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquired lock "refresh_cache-4b2fda93-a60d-4050-ae05-7a37ac284917" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.217455] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 609.241033] env[62521]: INFO nova.compute.manager [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Took 35.95 seconds to build instance. [ 609.421187] env[62521]: DEBUG nova.compute.manager [None req-dc1b6184-e5dc-4c96-a5d4-9986db2a0cd6 tempest-ServerDiagnosticsV248Test-347046341 tempest-ServerDiagnosticsV248Test-347046341-project-admin] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 609.422025] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411632cc-ceaf-4f66-a65a-522719ed665b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.430834] env[62521]: INFO nova.compute.manager [None req-dc1b6184-e5dc-4c96-a5d4-9986db2a0cd6 tempest-ServerDiagnosticsV248Test-347046341 tempest-ServerDiagnosticsV248Test-347046341-project-admin] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Retrieving diagnostics [ 609.431985] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893b29dc-64bb-49e9-8874-0beb35713452 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.544652] env[62521]: DEBUG oslo_vmware.api [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318409, 'name': PowerOnVM_Task, 'duration_secs': 0.444363} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.545369] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 609.545760] env[62521]: DEBUG nova.compute.manager [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 609.546725] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbcb09b-b214-4dcf-b4be-d1f407efd551 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.589037] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.595019] env[62521]: DEBUG oslo_concurrency.lockutils [None req-31deee70-46a2-499d-bdd9-f1b0865b4f08 tempest-ServersAdmin275Test-1000760649 tempest-ServersAdmin275Test-1000760649-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.595019] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.723s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.595019] env[62521]: DEBUG nova.objects.instance [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lazy-loading 'resources' on Instance uuid e976d2d2-aa69-473b-9dfc-e616b0132754 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 609.650365] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Successfully created port: 4ae88467-3176-442d-a60a-297ed80fae99 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.740704] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.743041] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7d3863f8-8bd8-4c95-8447-4354e0dea5a9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lock "7653f469-5ce5-4d74-9a3f-55a6adb3a255" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.715s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.908930] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.067493] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.247127] env[62521]: DEBUG nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 610.342720] env[62521]: DEBUG nova.compute.manager [req-5793bd59-5d35-4f3b-a1eb-c1523fd75963 req-a0a540ee-a9d2-4107-8be3-f0aa696a57d9 service nova] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Received event network-vif-deleted-7b451f88-f19d-4f24-9f30-a78410e4db4a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.412881] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Releasing lock "refresh_cache-4b2fda93-a60d-4050-ae05-7a37ac284917" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.413693] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.414035] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 610.414476] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-39c42ff2-5bad-4b4b-951d-5cc632e1fd1c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.428990] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11cd06a7-6f12-4120-b537-ddc829d2b7c1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.455730] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4b2fda93-a60d-4050-ae05-7a37ac284917 could not be found. [ 610.455972] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 610.456177] env[62521]: INFO nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Took 0.04 seconds to destroy the instance on the hypervisor. [ 610.456432] env[62521]: DEBUG oslo.service.loopingcall [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.459818] env[62521]: DEBUG nova.compute.manager [-] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.459818] env[62521]: DEBUG nova.network.neutron [-] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.479506] env[62521]: DEBUG nova.network.neutron [-] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.599998] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.641600] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.641600] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.641801] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.641875] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.642305] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.642530] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.642811] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.643027] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.643958] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.643958] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.643958] env[62521]: DEBUG nova.virt.hardware [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.644870] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be30f6a4-acf0-4590-9366-65430f308130 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.657451] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41347777-a483-43b8-b442-af91f6ddad75 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.687870] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a41435-245a-478d-813e-1d4a46493e11 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.698217] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e15441-0490-4212-873a-51e885b70fff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.730541] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a59c942-579c-4019-8d92-75ebdcb6d145 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.738348] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da81fae-49f2-49f2-8b19-a31fb13bf751 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.753610] env[62521]: DEBUG nova.compute.provider_tree [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.785893] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.970175] env[62521]: ERROR nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4ae88467-3176-442d-a60a-297ed80fae99, please check neutron logs for more information. [ 610.970175] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.970175] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.970175] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.970175] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.970175] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.970175] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.970175] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.970175] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.970175] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 610.970175] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.970175] env[62521]: ERROR nova.compute.manager raise self.value [ 610.970175] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.970175] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.970175] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.970175] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.970703] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.970703] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.970703] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4ae88467-3176-442d-a60a-297ed80fae99, please check neutron logs for more information. [ 610.970703] env[62521]: ERROR nova.compute.manager [ 610.970703] env[62521]: Traceback (most recent call last): [ 610.970703] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.970703] env[62521]: listener.cb(fileno) [ 610.970703] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.970703] env[62521]: result = function(*args, **kwargs) [ 610.970703] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.970703] env[62521]: return func(*args, **kwargs) [ 610.970703] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.970703] env[62521]: raise e [ 610.970703] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.970703] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 610.970703] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.970703] env[62521]: created_port_ids = self._update_ports_for_instance( [ 610.970703] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.970703] env[62521]: with excutils.save_and_reraise_exception(): [ 610.970703] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.970703] env[62521]: self.force_reraise() [ 610.970703] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.970703] env[62521]: raise self.value [ 610.970703] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.970703] env[62521]: updated_port = self._update_port( [ 610.970703] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.970703] env[62521]: _ensure_no_port_binding_failure(port) [ 610.970703] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.970703] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.971523] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 4ae88467-3176-442d-a60a-297ed80fae99, please check neutron logs for more information. [ 610.971523] env[62521]: Removing descriptor: 18 [ 610.971523] env[62521]: ERROR nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4ae88467-3176-442d-a60a-297ed80fae99, please check neutron logs for more information. [ 610.971523] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Traceback (most recent call last): [ 610.971523] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.971523] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] yield resources [ 610.971523] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.971523] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] self.driver.spawn(context, instance, image_meta, [ 610.971523] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.971523] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.971523] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.971523] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] vm_ref = self.build_virtual_machine(instance, [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] for vif in network_info: [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] return self._sync_wrapper(fn, *args, **kwargs) [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] self.wait() [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] self[:] = self._gt.wait() [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] return self._exit_event.wait() [ 610.971871] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] result = hub.switch() [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] return self.greenlet.switch() [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] result = function(*args, **kwargs) [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] return func(*args, **kwargs) [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] raise e [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] nwinfo = self.network_api.allocate_for_instance( [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.972405] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] created_port_ids = self._update_ports_for_instance( [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] with excutils.save_and_reraise_exception(): [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] self.force_reraise() [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] raise self.value [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] updated_port = self._update_port( [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] _ensure_no_port_binding_failure(port) [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.972808] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] raise exception.PortBindingFailed(port_id=port['id']) [ 610.973437] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] nova.exception.PortBindingFailed: Binding failed for port 4ae88467-3176-442d-a60a-297ed80fae99, please check neutron logs for more information. [ 610.973437] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] [ 610.973437] env[62521]: INFO nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Terminating instance [ 610.973545] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "refresh_cache-1d25cdca-3c87-47a2-85aa-86f7132297b5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.973622] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquired lock "refresh_cache-1d25cdca-3c87-47a2-85aa-86f7132297b5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.973790] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.986702] env[62521]: DEBUG nova.network.neutron [-] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.259644] env[62521]: DEBUG nova.scheduler.client.report [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.266069] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "20b01df0-4ced-4668-a786-046251a0c7ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.269358] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "20b01df0-4ced-4668-a786-046251a0c7ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.269635] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "20b01df0-4ced-4668-a786-046251a0c7ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.272549] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "20b01df0-4ced-4668-a786-046251a0c7ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.272549] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "20b01df0-4ced-4668-a786-046251a0c7ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.002s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.274975] env[62521]: INFO nova.compute.manager [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Terminating instance [ 611.276937] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "refresh_cache-20b01df0-4ced-4668-a786-046251a0c7ed" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.277434] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquired lock "refresh_cache-20b01df0-4ced-4668-a786-046251a0c7ed" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.277434] env[62521]: DEBUG nova.network.neutron [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.489470] env[62521]: INFO nova.compute.manager [-] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Took 1.03 seconds to deallocate network for instance. [ 611.492036] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.499204] env[62521]: DEBUG nova.compute.claims [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 611.499402] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.581200] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.768019] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.172s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.769344] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.232s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.769582] env[62521]: DEBUG nova.objects.instance [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lazy-loading 'resources' on Instance uuid 4f3a0a33-43c0-42fb-8660-198ff4a8529a {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 611.798079] env[62521]: INFO nova.scheduler.client.report [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Deleted allocations for instance e976d2d2-aa69-473b-9dfc-e616b0132754 [ 611.808428] env[62521]: DEBUG nova.network.neutron [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.905107] env[62521]: DEBUG nova.network.neutron [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.085164] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Releasing lock "refresh_cache-1d25cdca-3c87-47a2-85aa-86f7132297b5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.085164] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.085164] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.085164] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ae8cca1-5532-4b48-8bb0-1ef8a782a874 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.093853] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c13d0d-22e0-4257-915b-f6c6a2ae66c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.122542] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1d25cdca-3c87-47a2-85aa-86f7132297b5 could not be found. [ 612.122684] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 612.122862] env[62521]: INFO nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 612.123213] env[62521]: DEBUG oslo.service.loopingcall [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.123344] env[62521]: DEBUG nova.compute.manager [-] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.123449] env[62521]: DEBUG nova.network.neutron [-] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.145927] env[62521]: DEBUG nova.network.neutron [-] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.310634] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d1023d29-cfe1-4a81-a176-3ec20132b751 tempest-ServersAaction247Test-579727705 tempest-ServersAaction247Test-579727705-project-member] Lock "e976d2d2-aa69-473b-9dfc-e616b0132754" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.392s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.392398] env[62521]: DEBUG nova.compute.manager [req-c7c8627b-0de9-4af0-ad02-8148c6bb8f4c req-e9e2610b-dbc4-4f6f-808b-9ec095b93e17 service nova] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Received event network-changed-4ae88467-3176-442d-a60a-297ed80fae99 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.392398] env[62521]: DEBUG nova.compute.manager [req-c7c8627b-0de9-4af0-ad02-8148c6bb8f4c req-e9e2610b-dbc4-4f6f-808b-9ec095b93e17 service nova] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Refreshing instance network info cache due to event network-changed-4ae88467-3176-442d-a60a-297ed80fae99. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 612.392398] env[62521]: DEBUG oslo_concurrency.lockutils [req-c7c8627b-0de9-4af0-ad02-8148c6bb8f4c req-e9e2610b-dbc4-4f6f-808b-9ec095b93e17 service nova] Acquiring lock "refresh_cache-1d25cdca-3c87-47a2-85aa-86f7132297b5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.392745] env[62521]: DEBUG oslo_concurrency.lockutils [req-c7c8627b-0de9-4af0-ad02-8148c6bb8f4c req-e9e2610b-dbc4-4f6f-808b-9ec095b93e17 service nova] Acquired lock "refresh_cache-1d25cdca-3c87-47a2-85aa-86f7132297b5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.392990] env[62521]: DEBUG nova.network.neutron [req-c7c8627b-0de9-4af0-ad02-8148c6bb8f4c req-e9e2610b-dbc4-4f6f-808b-9ec095b93e17 service nova] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Refreshing network info cache for port 4ae88467-3176-442d-a60a-297ed80fae99 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 612.408015] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Releasing lock "refresh_cache-20b01df0-4ced-4668-a786-046251a0c7ed" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.408438] env[62521]: DEBUG nova.compute.manager [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.408628] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.409700] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b441e02b-02af-4cea-9f0a-c3fc5c41659f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.420164] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 612.420359] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f1bc086-7c47-4904-bf1d-0a18b22194f6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.428891] env[62521]: DEBUG oslo_vmware.api [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 612.428891] env[62521]: value = "task-1318410" [ 612.428891] env[62521]: _type = "Task" [ 612.428891] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.437985] env[62521]: DEBUG oslo_vmware.api [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318410, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.648066] env[62521]: DEBUG nova.network.neutron [-] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.770584] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7fabc0-f616-4613-af55-1c3ae1fe5870 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.779720] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c27b76-5ba4-4b29-ba43-a21dedb393af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.811682] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b67f2b-c52c-4e78-8430-769b193e2959 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.819473] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36964d48-cf86-4c05-93f5-986af8621379 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.833373] env[62521]: DEBUG nova.compute.provider_tree [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.917832] env[62521]: DEBUG nova.network.neutron [req-c7c8627b-0de9-4af0-ad02-8148c6bb8f4c req-e9e2610b-dbc4-4f6f-808b-9ec095b93e17 service nova] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.944750] env[62521]: DEBUG oslo_vmware.api [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318410, 'name': PowerOffVM_Task, 'duration_secs': 0.11825} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.945035] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 612.945232] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 612.945500] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e364f199-2554-42c0-bc72-4968a1998a8a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.969834] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 612.970065] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 612.970248] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Deleting the datastore file [datastore1] 20b01df0-4ced-4668-a786-046251a0c7ed {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 612.970511] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-76050262-acf2-4980-a277-f34163f3e951 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.977545] env[62521]: DEBUG oslo_vmware.api [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for the task: (returnval){ [ 612.977545] env[62521]: value = "task-1318412" [ 612.977545] env[62521]: _type = "Task" [ 612.977545] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.985230] env[62521]: DEBUG oslo_vmware.api [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318412, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.033176] env[62521]: DEBUG nova.network.neutron [req-c7c8627b-0de9-4af0-ad02-8148c6bb8f4c req-e9e2610b-dbc4-4f6f-808b-9ec095b93e17 service nova] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.153969] env[62521]: INFO nova.compute.manager [-] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Took 1.03 seconds to deallocate network for instance. [ 613.160027] env[62521]: DEBUG nova.compute.claims [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 613.161148] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.339063] env[62521]: DEBUG nova.scheduler.client.report [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.380165] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Acquiring lock "277ecbd7-e770-4ea0-ae2e-b49598428d0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.380686] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Lock "277ecbd7-e770-4ea0-ae2e-b49598428d0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.490703] env[62521]: DEBUG oslo_vmware.api [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Task: {'id': task-1318412, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139151} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.491603] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 613.491812] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 613.491993] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.492211] env[62521]: INFO nova.compute.manager [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Took 1.08 seconds to destroy the instance on the hypervisor. [ 613.492481] env[62521]: DEBUG oslo.service.loopingcall [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.492679] env[62521]: DEBUG nova.compute.manager [-] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.492787] env[62521]: DEBUG nova.network.neutron [-] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.510194] env[62521]: DEBUG nova.network.neutron [-] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.536165] env[62521]: DEBUG oslo_concurrency.lockutils [req-c7c8627b-0de9-4af0-ad02-8148c6bb8f4c req-e9e2610b-dbc4-4f6f-808b-9ec095b93e17 service nova] Releasing lock "refresh_cache-1d25cdca-3c87-47a2-85aa-86f7132297b5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.536439] env[62521]: DEBUG nova.compute.manager [req-c7c8627b-0de9-4af0-ad02-8148c6bb8f4c req-e9e2610b-dbc4-4f6f-808b-9ec095b93e17 service nova] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Received event network-vif-deleted-4ae88467-3176-442d-a60a-297ed80fae99 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.844327] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.075s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.846636] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.903s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.848291] env[62521]: INFO nova.compute.claims [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.866719] env[62521]: INFO nova.scheduler.client.report [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Deleted allocations for instance 4f3a0a33-43c0-42fb-8660-198ff4a8529a [ 614.013327] env[62521]: DEBUG nova.network.neutron [-] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.018208] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Acquiring lock "0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.018496] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Lock "0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.374990] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26fb24de-866e-4a22-aeba-ea9b282c56a4 tempest-ServersAdmin275Test-706599139 tempest-ServersAdmin275Test-706599139-project-member] Lock "4f3a0a33-43c0-42fb-8660-198ff4a8529a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.698s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.518156] env[62521]: INFO nova.compute.manager [-] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Took 1.03 seconds to deallocate network for instance. [ 615.033425] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.416373] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55cb221-4eee-4b09-9a58-d02afbdee02f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.424880] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2dfac7e-e51f-4477-ba91-a023da541ec2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.457571] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b389da26-8a56-467f-8975-91073e0727fc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.465575] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45909bd9-294d-4211-bb2e-03bc03eef8b1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.480227] env[62521]: DEBUG nova.compute.provider_tree [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.983270] env[62521]: DEBUG nova.scheduler.client.report [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.493743] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.647s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.494320] env[62521]: DEBUG nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.496955] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.528s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.498665] env[62521]: INFO nova.compute.claims [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 617.006949] env[62521]: DEBUG nova.compute.utils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.011547] env[62521]: DEBUG nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.012350] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.068010] env[62521]: DEBUG nova.policy [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd1fd76f1323e48f88490eac8cc96b2eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd670164b9d524f15911efc1681ac4bce', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 617.331911] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Successfully created port: d218a569-f06e-4f89-934c-288f23df0e46 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.516307] env[62521]: DEBUG nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 617.803951] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Acquiring lock "32c0268b-eee6-4e45-b9c3-077ec81bd314" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.804393] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Lock "32c0268b-eee6-4e45-b9c3-077ec81bd314" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.013629] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6b5985-11a9-4387-983a-c15696396f0c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.027581] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81aee50a-d4e3-4f07-9d1b-ebf750aefdb1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.057735] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa58561e-39f7-421b-94c9-ebc88e35809f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.065306] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76127387-f0ea-4fec-ac3f-77756dc1e571 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.079400] env[62521]: DEBUG nova.compute.provider_tree [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.172940] env[62521]: DEBUG nova.compute.manager [req-7adcd53c-9d33-435e-a3d1-eb3b5951dddb req-0663159d-18e6-415f-baa3-893923919f27 service nova] [instance: 1c892e78-a863-4561-940d-b44701e97041] Received event network-changed-d218a569-f06e-4f89-934c-288f23df0e46 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.173463] env[62521]: DEBUG nova.compute.manager [req-7adcd53c-9d33-435e-a3d1-eb3b5951dddb req-0663159d-18e6-415f-baa3-893923919f27 service nova] [instance: 1c892e78-a863-4561-940d-b44701e97041] Refreshing instance network info cache due to event network-changed-d218a569-f06e-4f89-934c-288f23df0e46. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 618.173691] env[62521]: DEBUG oslo_concurrency.lockutils [req-7adcd53c-9d33-435e-a3d1-eb3b5951dddb req-0663159d-18e6-415f-baa3-893923919f27 service nova] Acquiring lock "refresh_cache-1c892e78-a863-4561-940d-b44701e97041" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.173846] env[62521]: DEBUG oslo_concurrency.lockutils [req-7adcd53c-9d33-435e-a3d1-eb3b5951dddb req-0663159d-18e6-415f-baa3-893923919f27 service nova] Acquired lock "refresh_cache-1c892e78-a863-4561-940d-b44701e97041" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.175098] env[62521]: DEBUG nova.network.neutron [req-7adcd53c-9d33-435e-a3d1-eb3b5951dddb req-0663159d-18e6-415f-baa3-893923919f27 service nova] [instance: 1c892e78-a863-4561-940d-b44701e97041] Refreshing network info cache for port d218a569-f06e-4f89-934c-288f23df0e46 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.345012] env[62521]: ERROR nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d218a569-f06e-4f89-934c-288f23df0e46, please check neutron logs for more information. [ 618.345012] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.345012] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.345012] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.345012] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.345012] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.345012] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.345012] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.345012] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.345012] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 618.345012] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.345012] env[62521]: ERROR nova.compute.manager raise self.value [ 618.345012] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.345012] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.345012] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.345012] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.345472] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.345472] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.345472] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d218a569-f06e-4f89-934c-288f23df0e46, please check neutron logs for more information. [ 618.345472] env[62521]: ERROR nova.compute.manager [ 618.345472] env[62521]: Traceback (most recent call last): [ 618.345472] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.345472] env[62521]: listener.cb(fileno) [ 618.345472] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.345472] env[62521]: result = function(*args, **kwargs) [ 618.345472] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.345472] env[62521]: return func(*args, **kwargs) [ 618.345472] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.345472] env[62521]: raise e [ 618.345472] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.345472] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 618.345472] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.345472] env[62521]: created_port_ids = self._update_ports_for_instance( [ 618.345472] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.345472] env[62521]: with excutils.save_and_reraise_exception(): [ 618.345472] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.345472] env[62521]: self.force_reraise() [ 618.345472] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.345472] env[62521]: raise self.value [ 618.345472] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.345472] env[62521]: updated_port = self._update_port( [ 618.345472] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.345472] env[62521]: _ensure_no_port_binding_failure(port) [ 618.345472] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.345472] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.346405] env[62521]: nova.exception.PortBindingFailed: Binding failed for port d218a569-f06e-4f89-934c-288f23df0e46, please check neutron logs for more information. [ 618.346405] env[62521]: Removing descriptor: 18 [ 618.532350] env[62521]: DEBUG nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.557312] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.557573] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.557729] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.557910] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.558065] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.558212] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.558413] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.558568] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.558728] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.558886] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.559064] env[62521]: DEBUG nova.virt.hardware [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.559906] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6ee964-aa17-4a72-949f-6597216deb07 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.567343] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10fa65cf-06b2-4b23-890f-4c97ebdaa1cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.581808] env[62521]: ERROR nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d218a569-f06e-4f89-934c-288f23df0e46, please check neutron logs for more information. [ 618.581808] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] Traceback (most recent call last): [ 618.581808] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.581808] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] yield resources [ 618.581808] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.581808] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] self.driver.spawn(context, instance, image_meta, [ 618.581808] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 618.581808] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.581808] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.581808] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] vm_ref = self.build_virtual_machine(instance, [ 618.581808] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] for vif in network_info: [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] return self._sync_wrapper(fn, *args, **kwargs) [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] self.wait() [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] self[:] = self._gt.wait() [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] return self._exit_event.wait() [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 618.582222] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] current.throw(*self._exc) [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] result = function(*args, **kwargs) [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] return func(*args, **kwargs) [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] raise e [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] nwinfo = self.network_api.allocate_for_instance( [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] created_port_ids = self._update_ports_for_instance( [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] with excutils.save_and_reraise_exception(): [ 618.582484] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] self.force_reraise() [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] raise self.value [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] updated_port = self._update_port( [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] _ensure_no_port_binding_failure(port) [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] raise exception.PortBindingFailed(port_id=port['id']) [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] nova.exception.PortBindingFailed: Binding failed for port d218a569-f06e-4f89-934c-288f23df0e46, please check neutron logs for more information. [ 618.582803] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] [ 618.582803] env[62521]: INFO nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Terminating instance [ 618.584161] env[62521]: DEBUG nova.scheduler.client.report [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.587274] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquiring lock "refresh_cache-1c892e78-a863-4561-940d-b44701e97041" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.691671] env[62521]: DEBUG nova.network.neutron [req-7adcd53c-9d33-435e-a3d1-eb3b5951dddb req-0663159d-18e6-415f-baa3-893923919f27 service nova] [instance: 1c892e78-a863-4561-940d-b44701e97041] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.737618] env[62521]: DEBUG nova.network.neutron [req-7adcd53c-9d33-435e-a3d1-eb3b5951dddb req-0663159d-18e6-415f-baa3-893923919f27 service nova] [instance: 1c892e78-a863-4561-940d-b44701e97041] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.089172] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.592s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.089656] env[62521]: DEBUG nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.092671] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.524s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.093866] env[62521]: INFO nova.compute.claims [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.240490] env[62521]: DEBUG oslo_concurrency.lockutils [req-7adcd53c-9d33-435e-a3d1-eb3b5951dddb req-0663159d-18e6-415f-baa3-893923919f27 service nova] Releasing lock "refresh_cache-1c892e78-a863-4561-940d-b44701e97041" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.240916] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquired lock "refresh_cache-1c892e78-a863-4561-940d-b44701e97041" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.241109] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.599181] env[62521]: DEBUG nova.compute.utils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.599945] env[62521]: DEBUG nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 619.600032] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 619.667290] env[62521]: DEBUG nova.policy [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '252852096a844d67b9afe3dd1f68fbda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9bdaf26dd2de43de94d707621d6dba75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 619.759832] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.819849] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.980673] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Successfully created port: 239afdcf-1140-44fd-9280-cfb1580f845c {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.022042] env[62521]: DEBUG nova.compute.manager [None req-15902866-ca4c-401f-99f6-dac55bbc3341 tempest-ServerDiagnosticsV248Test-347046341 tempest-ServerDiagnosticsV248Test-347046341-project-admin] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 620.023552] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984854cb-8853-4f10-9749-ffab3ef0c0a4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.030527] env[62521]: INFO nova.compute.manager [None req-15902866-ca4c-401f-99f6-dac55bbc3341 tempest-ServerDiagnosticsV248Test-347046341 tempest-ServerDiagnosticsV248Test-347046341-project-admin] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Retrieving diagnostics [ 620.031452] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3d8e20-9859-4575-8d8a-0a5b7f1ee892 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.109191] env[62521]: DEBUG nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.224848] env[62521]: DEBUG nova.compute.manager [req-fb0b68ac-2dcb-4848-8d0f-a35a18600219 req-152f6f5c-9324-4c76-a5fa-8fddfa3ecbed service nova] [instance: 1c892e78-a863-4561-940d-b44701e97041] Received event network-vif-deleted-d218a569-f06e-4f89-934c-288f23df0e46 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.321829] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Releasing lock "refresh_cache-1c892e78-a863-4561-940d-b44701e97041" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.322324] env[62521]: DEBUG nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.322461] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.322701] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b70611e0-945d-46b3-9d8a-c12943936043 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.334436] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e254eb8-c622-4659-84f7-ce53a3e4e13d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.359094] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1c892e78-a863-4561-940d-b44701e97041 could not be found. [ 620.359330] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.359512] env[62521]: INFO nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Took 0.04 seconds to destroy the instance on the hypervisor. [ 620.359754] env[62521]: DEBUG oslo.service.loopingcall [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.362110] env[62521]: DEBUG nova.compute.manager [-] [instance: 1c892e78-a863-4561-940d-b44701e97041] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.362215] env[62521]: DEBUG nova.network.neutron [-] [instance: 1c892e78-a863-4561-940d-b44701e97041] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.384535] env[62521]: DEBUG nova.network.neutron [-] [instance: 1c892e78-a863-4561-940d-b44701e97041] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.581014] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dcb21f0-1146-4933-8e26-bcb6db290e6f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.590211] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b084b8-13fb-4c83-a9f2-df18a92f8b67 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.625052] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bd500d-4ceb-40b8-bcdd-e4d4c1711839 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.632977] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43be576a-ff49-4725-9aa6-f0bf153cf95e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.648653] env[62521]: DEBUG nova.compute.provider_tree [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.889460] env[62521]: DEBUG nova.network.neutron [-] [instance: 1c892e78-a863-4561-940d-b44701e97041] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.129259] env[62521]: DEBUG nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.132638] env[62521]: ERROR nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 239afdcf-1140-44fd-9280-cfb1580f845c, please check neutron logs for more information. [ 621.132638] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 621.132638] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.132638] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 621.132638] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.132638] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 621.132638] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.132638] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 621.132638] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.132638] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 621.132638] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.132638] env[62521]: ERROR nova.compute.manager raise self.value [ 621.132638] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.132638] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 621.132638] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.132638] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 621.133022] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.133022] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 621.133022] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 239afdcf-1140-44fd-9280-cfb1580f845c, please check neutron logs for more information. [ 621.133022] env[62521]: ERROR nova.compute.manager [ 621.133022] env[62521]: Traceback (most recent call last): [ 621.133022] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 621.133022] env[62521]: listener.cb(fileno) [ 621.133022] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.133022] env[62521]: result = function(*args, **kwargs) [ 621.133022] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.133022] env[62521]: return func(*args, **kwargs) [ 621.133022] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.133022] env[62521]: raise e [ 621.133022] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.133022] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 621.133022] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.133022] env[62521]: created_port_ids = self._update_ports_for_instance( [ 621.133022] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.133022] env[62521]: with excutils.save_and_reraise_exception(): [ 621.133022] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.133022] env[62521]: self.force_reraise() [ 621.133022] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.133022] env[62521]: raise self.value [ 621.133022] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.133022] env[62521]: updated_port = self._update_port( [ 621.133022] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.133022] env[62521]: _ensure_no_port_binding_failure(port) [ 621.133022] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.133022] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 621.133793] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 239afdcf-1140-44fd-9280-cfb1580f845c, please check neutron logs for more information. [ 621.133793] env[62521]: Removing descriptor: 18 [ 621.152478] env[62521]: DEBUG nova.scheduler.client.report [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.163057] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquiring lock "7653f469-5ce5-4d74-9a3f-55a6adb3a255" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.163372] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lock "7653f469-5ce5-4d74-9a3f-55a6adb3a255" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.163622] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquiring lock "7653f469-5ce5-4d74-9a3f-55a6adb3a255-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.163844] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lock "7653f469-5ce5-4d74-9a3f-55a6adb3a255-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.164062] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lock "7653f469-5ce5-4d74-9a3f-55a6adb3a255-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.168616] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.168878] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.169084] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.169313] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.169496] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.169705] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.169946] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.171052] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.171052] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.171052] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.172251] env[62521]: DEBUG nova.virt.hardware [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.175603] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0794f95f-37fd-4433-8021-f6859bd4c99d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.178931] env[62521]: INFO nova.compute.manager [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Terminating instance [ 621.182029] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquiring lock "refresh_cache-7653f469-5ce5-4d74-9a3f-55a6adb3a255" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.182293] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquired lock "refresh_cache-7653f469-5ce5-4d74-9a3f-55a6adb3a255" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.182509] env[62521]: DEBUG nova.network.neutron [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.189522] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d83b07-d019-43ed-83c5-5868fff86ba8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.204469] env[62521]: ERROR nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 239afdcf-1140-44fd-9280-cfb1580f845c, please check neutron logs for more information. [ 621.204469] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Traceback (most recent call last): [ 621.204469] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 621.204469] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] yield resources [ 621.204469] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.204469] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] self.driver.spawn(context, instance, image_meta, [ 621.204469] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 621.204469] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.204469] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.204469] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] vm_ref = self.build_virtual_machine(instance, [ 621.204469] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] for vif in network_info: [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] return self._sync_wrapper(fn, *args, **kwargs) [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] self.wait() [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] self[:] = self._gt.wait() [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] return self._exit_event.wait() [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 621.204792] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] current.throw(*self._exc) [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] result = function(*args, **kwargs) [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] return func(*args, **kwargs) [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] raise e [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] nwinfo = self.network_api.allocate_for_instance( [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] created_port_ids = self._update_ports_for_instance( [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] with excutils.save_and_reraise_exception(): [ 621.205100] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] self.force_reraise() [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] raise self.value [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] updated_port = self._update_port( [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] _ensure_no_port_binding_failure(port) [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] raise exception.PortBindingFailed(port_id=port['id']) [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] nova.exception.PortBindingFailed: Binding failed for port 239afdcf-1140-44fd-9280-cfb1580f845c, please check neutron logs for more information. [ 621.205402] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] [ 621.205402] env[62521]: INFO nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Terminating instance [ 621.207359] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Acquiring lock "refresh_cache-9c2c0e66-5c67-4e0a-9205-4080bda44a66" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.207567] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Acquired lock "refresh_cache-9c2c0e66-5c67-4e0a-9205-4080bda44a66" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.207772] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.391886] env[62521]: INFO nova.compute.manager [-] [instance: 1c892e78-a863-4561-940d-b44701e97041] Took 1.03 seconds to deallocate network for instance. [ 621.394686] env[62521]: DEBUG nova.compute.claims [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.394862] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.659128] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.659678] env[62521]: DEBUG nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.666035] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.508s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.667598] env[62521]: INFO nova.compute.claims [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.703356] env[62521]: DEBUG nova.network.neutron [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.732750] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.776839] env[62521]: DEBUG nova.network.neutron [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.897770] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.176259] env[62521]: DEBUG nova.compute.utils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.181027] env[62521]: DEBUG nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 622.181027] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 622.244461] env[62521]: DEBUG nova.policy [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e3eb9ead5aa44410af02786a42a6b6cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '92529cdd47444a3681f5f05ec22983e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 622.281357] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Releasing lock "refresh_cache-7653f469-5ce5-4d74-9a3f-55a6adb3a255" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.281779] env[62521]: DEBUG nova.compute.manager [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.281969] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.282886] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2f2fff-2175-4602-8d2d-d7205abcfe27 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.290569] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 622.292636] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bc5be29-b20d-4835-84fd-c3fb9ed33aab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.299147] env[62521]: DEBUG oslo_vmware.api [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for the task: (returnval){ [ 622.299147] env[62521]: value = "task-1318413" [ 622.299147] env[62521]: _type = "Task" [ 622.299147] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.306836] env[62521]: DEBUG oslo_vmware.api [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318413, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.402124] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Releasing lock "refresh_cache-9c2c0e66-5c67-4e0a-9205-4080bda44a66" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.402475] env[62521]: DEBUG nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.402641] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.402958] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06ae1aeb-e2fd-42b1-ae1f-dc379461ef40 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.417501] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f809db0-7208-48eb-8647-1255f69eb2e6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.430967] env[62521]: DEBUG nova.compute.manager [req-468104bb-882f-4c47-bb62-0322a2f9c3dc req-1fdf19d1-909a-4b03-a27f-7f6939ff701e service nova] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Received event network-changed-239afdcf-1140-44fd-9280-cfb1580f845c {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.430967] env[62521]: DEBUG nova.compute.manager [req-468104bb-882f-4c47-bb62-0322a2f9c3dc req-1fdf19d1-909a-4b03-a27f-7f6939ff701e service nova] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Refreshing instance network info cache due to event network-changed-239afdcf-1140-44fd-9280-cfb1580f845c. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 622.432723] env[62521]: DEBUG oslo_concurrency.lockutils [req-468104bb-882f-4c47-bb62-0322a2f9c3dc req-1fdf19d1-909a-4b03-a27f-7f6939ff701e service nova] Acquiring lock "refresh_cache-9c2c0e66-5c67-4e0a-9205-4080bda44a66" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.432723] env[62521]: DEBUG oslo_concurrency.lockutils [req-468104bb-882f-4c47-bb62-0322a2f9c3dc req-1fdf19d1-909a-4b03-a27f-7f6939ff701e service nova] Acquired lock "refresh_cache-9c2c0e66-5c67-4e0a-9205-4080bda44a66" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.432723] env[62521]: DEBUG nova.network.neutron [req-468104bb-882f-4c47-bb62-0322a2f9c3dc req-1fdf19d1-909a-4b03-a27f-7f6939ff701e service nova] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Refreshing network info cache for port 239afdcf-1140-44fd-9280-cfb1580f845c {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 622.445892] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9c2c0e66-5c67-4e0a-9205-4080bda44a66 could not be found. [ 622.446135] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.446319] env[62521]: INFO nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Took 0.04 seconds to destroy the instance on the hypervisor. [ 622.446570] env[62521]: DEBUG oslo.service.loopingcall [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.447358] env[62521]: DEBUG nova.compute.manager [-] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.447463] env[62521]: DEBUG nova.network.neutron [-] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.472838] env[62521]: DEBUG nova.network.neutron [-] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.681326] env[62521]: DEBUG nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 622.716693] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Successfully created port: 1f30572e-7dc3-4725-943c-feab59bb48ef {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.814704] env[62521]: DEBUG oslo_vmware.api [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318413, 'name': PowerOffVM_Task, 'duration_secs': 0.123145} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.815289] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 622.815963] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 622.818985] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a9417dcb-5067-46cb-a5ec-3a5436c74af9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.847195] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 622.847430] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 622.847614] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Deleting the datastore file [datastore1] 7653f469-5ce5-4d74-9a3f-55a6adb3a255 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 622.847869] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ce6bfda-5b8f-42cd-8574-c9b2c11a271c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.856545] env[62521]: DEBUG oslo_vmware.api [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for the task: (returnval){ [ 622.856545] env[62521]: value = "task-1318415" [ 622.856545] env[62521]: _type = "Task" [ 622.856545] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.863875] env[62521]: DEBUG oslo_vmware.api [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318415, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.963505] env[62521]: DEBUG nova.network.neutron [req-468104bb-882f-4c47-bb62-0322a2f9c3dc req-1fdf19d1-909a-4b03-a27f-7f6939ff701e service nova] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.977413] env[62521]: DEBUG nova.network.neutron [-] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.008654] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Acquiring lock "63abbe0a-af36-4679-8cf1-87f9e9abf7b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.009500] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Lock "63abbe0a-af36-4679-8cf1-87f9e9abf7b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.051204] env[62521]: DEBUG nova.network.neutron [req-468104bb-882f-4c47-bb62-0322a2f9c3dc req-1fdf19d1-909a-4b03-a27f-7f6939ff701e service nova] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.185331] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86326e15-562c-49b9-9cad-11839334a164 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.199012] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b26253-95b3-4884-93d2-580489905917 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.229870] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b51a2f2-68c5-47f5-a2b7-43d59442731e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.238783] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d984c36-506e-46df-be5e-3ee89e79618e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.254710] env[62521]: DEBUG nova.compute.provider_tree [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.367504] env[62521]: DEBUG oslo_vmware.api [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Task: {'id': task-1318415, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.0986} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.367757] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 623.367940] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 623.370839] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 623.370884] env[62521]: INFO nova.compute.manager [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Took 1.09 seconds to destroy the instance on the hypervisor. [ 623.371784] env[62521]: DEBUG oslo.service.loopingcall [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.371784] env[62521]: DEBUG nova.compute.manager [-] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.371784] env[62521]: DEBUG nova.network.neutron [-] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.390549] env[62521]: DEBUG nova.network.neutron [-] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.480217] env[62521]: INFO nova.compute.manager [-] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Took 1.03 seconds to deallocate network for instance. [ 623.485586] env[62521]: DEBUG nova.compute.claims [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.485816] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.553741] env[62521]: DEBUG oslo_concurrency.lockutils [req-468104bb-882f-4c47-bb62-0322a2f9c3dc req-1fdf19d1-909a-4b03-a27f-7f6939ff701e service nova] Releasing lock "refresh_cache-9c2c0e66-5c67-4e0a-9205-4080bda44a66" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.554071] env[62521]: DEBUG nova.compute.manager [req-468104bb-882f-4c47-bb62-0322a2f9c3dc req-1fdf19d1-909a-4b03-a27f-7f6939ff701e service nova] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Received event network-vif-deleted-239afdcf-1140-44fd-9280-cfb1580f845c {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 623.694403] env[62521]: DEBUG nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 623.737537] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.738447] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.738447] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.738447] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.738447] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.738610] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.738996] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.738996] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.739140] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.739301] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.739467] env[62521]: DEBUG nova.virt.hardware [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.740783] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d02384-edd3-44f1-a2db-e52873ac145f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.753032] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5130e1-c5c1-4ace-9f6a-ac7d172515b0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.757344] env[62521]: DEBUG nova.scheduler.client.report [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.855934] env[62521]: DEBUG nova.compute.manager [req-8b9c3e19-7b5c-489a-8e3c-1dedfdae1033 req-8729437f-c366-4afb-8dd6-e7299ec66b0e service nova] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Received event network-changed-1f30572e-7dc3-4725-943c-feab59bb48ef {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 623.856155] env[62521]: DEBUG nova.compute.manager [req-8b9c3e19-7b5c-489a-8e3c-1dedfdae1033 req-8729437f-c366-4afb-8dd6-e7299ec66b0e service nova] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Refreshing instance network info cache due to event network-changed-1f30572e-7dc3-4725-943c-feab59bb48ef. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 623.856362] env[62521]: DEBUG oslo_concurrency.lockutils [req-8b9c3e19-7b5c-489a-8e3c-1dedfdae1033 req-8729437f-c366-4afb-8dd6-e7299ec66b0e service nova] Acquiring lock "refresh_cache-da9d2279-79ae-4371-bef5-08ae15f2d9c9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.856505] env[62521]: DEBUG oslo_concurrency.lockutils [req-8b9c3e19-7b5c-489a-8e3c-1dedfdae1033 req-8729437f-c366-4afb-8dd6-e7299ec66b0e service nova] Acquired lock "refresh_cache-da9d2279-79ae-4371-bef5-08ae15f2d9c9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.856664] env[62521]: DEBUG nova.network.neutron [req-8b9c3e19-7b5c-489a-8e3c-1dedfdae1033 req-8729437f-c366-4afb-8dd6-e7299ec66b0e service nova] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Refreshing network info cache for port 1f30572e-7dc3-4725-943c-feab59bb48ef {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 623.896850] env[62521]: DEBUG nova.network.neutron [-] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.058221] env[62521]: ERROR nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f30572e-7dc3-4725-943c-feab59bb48ef, please check neutron logs for more information. [ 624.058221] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.058221] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.058221] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.058221] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.058221] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.058221] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.058221] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.058221] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.058221] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 624.058221] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.058221] env[62521]: ERROR nova.compute.manager raise self.value [ 624.058221] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.058221] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.058221] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.058221] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.058644] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.058644] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.058644] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f30572e-7dc3-4725-943c-feab59bb48ef, please check neutron logs for more information. [ 624.058644] env[62521]: ERROR nova.compute.manager [ 624.058644] env[62521]: Traceback (most recent call last): [ 624.058644] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.058644] env[62521]: listener.cb(fileno) [ 624.058644] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.058644] env[62521]: result = function(*args, **kwargs) [ 624.058644] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.058644] env[62521]: return func(*args, **kwargs) [ 624.058644] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.058644] env[62521]: raise e [ 624.058644] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.058644] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 624.058644] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.058644] env[62521]: created_port_ids = self._update_ports_for_instance( [ 624.058644] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.058644] env[62521]: with excutils.save_and_reraise_exception(): [ 624.058644] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.058644] env[62521]: self.force_reraise() [ 624.058644] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.058644] env[62521]: raise self.value [ 624.058644] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.058644] env[62521]: updated_port = self._update_port( [ 624.058644] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.058644] env[62521]: _ensure_no_port_binding_failure(port) [ 624.058644] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.058644] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.059359] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 1f30572e-7dc3-4725-943c-feab59bb48ef, please check neutron logs for more information. [ 624.059359] env[62521]: Removing descriptor: 15 [ 624.059359] env[62521]: ERROR nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f30572e-7dc3-4725-943c-feab59bb48ef, please check neutron logs for more information. [ 624.059359] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Traceback (most recent call last): [ 624.059359] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 624.059359] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] yield resources [ 624.059359] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.059359] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] self.driver.spawn(context, instance, image_meta, [ 624.059359] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 624.059359] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.059359] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.059359] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] vm_ref = self.build_virtual_machine(instance, [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] for vif in network_info: [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] return self._sync_wrapper(fn, *args, **kwargs) [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] self.wait() [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] self[:] = self._gt.wait() [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] return self._exit_event.wait() [ 624.059666] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] result = hub.switch() [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] return self.greenlet.switch() [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] result = function(*args, **kwargs) [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] return func(*args, **kwargs) [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] raise e [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] nwinfo = self.network_api.allocate_for_instance( [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.059936] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] created_port_ids = self._update_ports_for_instance( [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] with excutils.save_and_reraise_exception(): [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] self.force_reraise() [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] raise self.value [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] updated_port = self._update_port( [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] _ensure_no_port_binding_failure(port) [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.060285] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] raise exception.PortBindingFailed(port_id=port['id']) [ 624.060583] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] nova.exception.PortBindingFailed: Binding failed for port 1f30572e-7dc3-4725-943c-feab59bb48ef, please check neutron logs for more information. [ 624.060583] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] [ 624.060583] env[62521]: INFO nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Terminating instance [ 624.063057] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Acquiring lock "refresh_cache-da9d2279-79ae-4371-bef5-08ae15f2d9c9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.272330] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.607s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.272978] env[62521]: DEBUG nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 624.275858] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.558s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.277883] env[62521]: INFO nova.compute.claims [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.380604] env[62521]: DEBUG nova.network.neutron [req-8b9c3e19-7b5c-489a-8e3c-1dedfdae1033 req-8729437f-c366-4afb-8dd6-e7299ec66b0e service nova] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.400306] env[62521]: INFO nova.compute.manager [-] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Took 1.03 seconds to deallocate network for instance. [ 624.460691] env[62521]: DEBUG nova.network.neutron [req-8b9c3e19-7b5c-489a-8e3c-1dedfdae1033 req-8729437f-c366-4afb-8dd6-e7299ec66b0e service nova] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.782605] env[62521]: DEBUG nova.compute.utils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.785587] env[62521]: DEBUG nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 624.785759] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 624.852400] env[62521]: DEBUG nova.policy [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0822dd896d5b4047b82d733dc40aab94', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1bc836b9837e47d5826f6c08f66bc813', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.907426] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.962832] env[62521]: DEBUG oslo_concurrency.lockutils [req-8b9c3e19-7b5c-489a-8e3c-1dedfdae1033 req-8729437f-c366-4afb-8dd6-e7299ec66b0e service nova] Releasing lock "refresh_cache-da9d2279-79ae-4371-bef5-08ae15f2d9c9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.963533] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Acquired lock "refresh_cache-da9d2279-79ae-4371-bef5-08ae15f2d9c9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.963731] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.214626] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Successfully created port: 1283c88b-3962-4229-abfd-93a41daf5b29 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 625.286344] env[62521]: DEBUG nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 625.490447] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.575871] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.733011] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eecc7886-452a-43b7-b624-7e90a1077b14 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.740930] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8422c36-3256-4414-b6b5-5498ffcd60df {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.773886] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5949163f-981e-4fe7-84ff-727e8bd04b3d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.781319] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db60687f-790c-4a16-9263-69877a6e1a42 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.799953] env[62521]: DEBUG nova.compute.provider_tree [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.879234] env[62521]: DEBUG nova.compute.manager [req-38f14015-dd06-49ce-99d2-f655e0ece779 req-555e17bb-2526-45b6-8866-f352457de463 service nova] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Received event network-vif-deleted-1f30572e-7dc3-4725-943c-feab59bb48ef {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.078726] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Releasing lock "refresh_cache-da9d2279-79ae-4371-bef5-08ae15f2d9c9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.079196] env[62521]: DEBUG nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 626.079397] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 626.079706] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b7e3705-5297-49c3-9531-9f3ab630f1b5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.090204] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb156784-bd38-4d7b-8dea-4ff6575031fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.116681] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance da9d2279-79ae-4371-bef5-08ae15f2d9c9 could not be found. [ 626.116905] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 626.117103] env[62521]: INFO nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 626.117350] env[62521]: DEBUG oslo.service.loopingcall [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.117565] env[62521]: DEBUG nova.compute.manager [-] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.117661] env[62521]: DEBUG nova.network.neutron [-] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.133074] env[62521]: DEBUG nova.network.neutron [-] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.302854] env[62521]: DEBUG nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.305575] env[62521]: DEBUG nova.scheduler.client.report [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.334312] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.334566] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.334725] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.334907] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.335072] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.335252] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.335499] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.335666] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.335834] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.335991] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.336178] env[62521]: DEBUG nova.virt.hardware [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.337012] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c37f155-6445-4916-b1ce-3f667371949b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.345384] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a413c59c-3ef2-4ba2-840c-d92e02a4b2df {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.393479] env[62521]: ERROR nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1283c88b-3962-4229-abfd-93a41daf5b29, please check neutron logs for more information. [ 626.393479] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.393479] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.393479] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.393479] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.393479] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.393479] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.393479] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.393479] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.393479] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 626.393479] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.393479] env[62521]: ERROR nova.compute.manager raise self.value [ 626.393479] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.393479] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.393479] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.393479] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.393970] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.393970] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.393970] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1283c88b-3962-4229-abfd-93a41daf5b29, please check neutron logs for more information. [ 626.393970] env[62521]: ERROR nova.compute.manager [ 626.393970] env[62521]: Traceback (most recent call last): [ 626.393970] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.393970] env[62521]: listener.cb(fileno) [ 626.393970] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.393970] env[62521]: result = function(*args, **kwargs) [ 626.393970] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.393970] env[62521]: return func(*args, **kwargs) [ 626.393970] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.393970] env[62521]: raise e [ 626.393970] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.393970] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 626.393970] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.393970] env[62521]: created_port_ids = self._update_ports_for_instance( [ 626.393970] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.393970] env[62521]: with excutils.save_and_reraise_exception(): [ 626.393970] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.393970] env[62521]: self.force_reraise() [ 626.393970] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.393970] env[62521]: raise self.value [ 626.393970] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.393970] env[62521]: updated_port = self._update_port( [ 626.393970] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.393970] env[62521]: _ensure_no_port_binding_failure(port) [ 626.393970] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.393970] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.394748] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 1283c88b-3962-4229-abfd-93a41daf5b29, please check neutron logs for more information. [ 626.394748] env[62521]: Removing descriptor: 15 [ 626.394748] env[62521]: ERROR nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1283c88b-3962-4229-abfd-93a41daf5b29, please check neutron logs for more information. [ 626.394748] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Traceback (most recent call last): [ 626.394748] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 626.394748] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] yield resources [ 626.394748] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.394748] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] self.driver.spawn(context, instance, image_meta, [ 626.394748] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 626.394748] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.394748] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.394748] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] vm_ref = self.build_virtual_machine(instance, [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] for vif in network_info: [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] return self._sync_wrapper(fn, *args, **kwargs) [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] self.wait() [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] self[:] = self._gt.wait() [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] return self._exit_event.wait() [ 626.395068] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] result = hub.switch() [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] return self.greenlet.switch() [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] result = function(*args, **kwargs) [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] return func(*args, **kwargs) [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] raise e [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] nwinfo = self.network_api.allocate_for_instance( [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.395413] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] created_port_ids = self._update_ports_for_instance( [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] with excutils.save_and_reraise_exception(): [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] self.force_reraise() [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] raise self.value [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] updated_port = self._update_port( [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] _ensure_no_port_binding_failure(port) [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.395737] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] raise exception.PortBindingFailed(port_id=port['id']) [ 626.396052] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] nova.exception.PortBindingFailed: Binding failed for port 1283c88b-3962-4229-abfd-93a41daf5b29, please check neutron logs for more information. [ 626.396052] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] [ 626.396052] env[62521]: INFO nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Terminating instance [ 626.397265] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Acquiring lock "refresh_cache-e944e6e5-7bc0-4efb-a811-38221734ec4f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.397425] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Acquired lock "refresh_cache-e944e6e5-7bc0-4efb-a811-38221734ec4f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.397589] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.635500] env[62521]: DEBUG nova.network.neutron [-] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.813154] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.813614] env[62521]: DEBUG nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 626.816167] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.938s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.914346] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.035718] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.138841] env[62521]: INFO nova.compute.manager [-] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Took 1.02 seconds to deallocate network for instance. [ 627.141160] env[62521]: DEBUG nova.compute.claims [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.141343] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.322117] env[62521]: DEBUG nova.compute.utils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.323971] env[62521]: DEBUG nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 627.323971] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 627.400448] env[62521]: DEBUG nova.policy [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf5192a2e45643abb7b24031114814be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f3e0e94b61c45aebda4e7038a1c9aed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 627.542193] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Releasing lock "refresh_cache-e944e6e5-7bc0-4efb-a811-38221734ec4f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.542697] env[62521]: DEBUG nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 627.542890] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 627.543692] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9efc51a2-3fb0-4180-bfe2-4370a9a0bf70 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.553069] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f41396e-5db8-4187-bc74-52174d2db7cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.577777] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e944e6e5-7bc0-4efb-a811-38221734ec4f could not be found. [ 627.578098] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 627.578437] env[62521]: INFO nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 627.578745] env[62521]: DEBUG oslo.service.loopingcall [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 627.581180] env[62521]: DEBUG nova.compute.manager [-] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.581339] env[62521]: DEBUG nova.network.neutron [-] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.599235] env[62521]: DEBUG nova.network.neutron [-] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.761251] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a688d0e0-fb33-460e-ada3-b7dc402a018b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.770176] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ca5842-8d73-4af5-826a-3dff72dcd0db {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.798870] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Successfully created port: 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 627.801312] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bae6b39-40a3-4605-9ae3-594a3dc5cfd3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.808864] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ab00c1-c105-47ec-8841-7736bdc2a515 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.824407] env[62521]: DEBUG nova.compute.provider_tree [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.829757] env[62521]: DEBUG nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 627.978068] env[62521]: DEBUG nova.compute.manager [req-76656e51-44cd-4f90-a3d5-21bcc8ecedc2 req-64056439-1b95-459d-a508-dd845a5dd64a service nova] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Received event network-changed-1283c88b-3962-4229-abfd-93a41daf5b29 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.978198] env[62521]: DEBUG nova.compute.manager [req-76656e51-44cd-4f90-a3d5-21bcc8ecedc2 req-64056439-1b95-459d-a508-dd845a5dd64a service nova] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Refreshing instance network info cache due to event network-changed-1283c88b-3962-4229-abfd-93a41daf5b29. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 627.978450] env[62521]: DEBUG oslo_concurrency.lockutils [req-76656e51-44cd-4f90-a3d5-21bcc8ecedc2 req-64056439-1b95-459d-a508-dd845a5dd64a service nova] Acquiring lock "refresh_cache-e944e6e5-7bc0-4efb-a811-38221734ec4f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.978550] env[62521]: DEBUG oslo_concurrency.lockutils [req-76656e51-44cd-4f90-a3d5-21bcc8ecedc2 req-64056439-1b95-459d-a508-dd845a5dd64a service nova] Acquired lock "refresh_cache-e944e6e5-7bc0-4efb-a811-38221734ec4f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.978707] env[62521]: DEBUG nova.network.neutron [req-76656e51-44cd-4f90-a3d5-21bcc8ecedc2 req-64056439-1b95-459d-a508-dd845a5dd64a service nova] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Refreshing network info cache for port 1283c88b-3962-4229-abfd-93a41daf5b29 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 628.102051] env[62521]: DEBUG nova.network.neutron [-] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.329978] env[62521]: DEBUG nova.scheduler.client.report [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.508091] env[62521]: DEBUG nova.network.neutron [req-76656e51-44cd-4f90-a3d5-21bcc8ecedc2 req-64056439-1b95-459d-a508-dd845a5dd64a service nova] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.587792] env[62521]: DEBUG nova.network.neutron [req-76656e51-44cd-4f90-a3d5-21bcc8ecedc2 req-64056439-1b95-459d-a508-dd845a5dd64a service nova] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.606958] env[62521]: INFO nova.compute.manager [-] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Took 1.03 seconds to deallocate network for instance. [ 628.609705] env[62521]: DEBUG nova.compute.claims [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 628.610068] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.833636] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.017s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.834355] env[62521]: ERROR nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5d67aeef-bad4-4299-bc25-6652e49a217f, please check neutron logs for more information. [ 628.834355] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Traceback (most recent call last): [ 628.834355] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.834355] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] self.driver.spawn(context, instance, image_meta, [ 628.834355] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 628.834355] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.834355] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.834355] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] vm_ref = self.build_virtual_machine(instance, [ 628.834355] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.834355] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.834355] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] for vif in network_info: [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] return self._sync_wrapper(fn, *args, **kwargs) [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] self.wait() [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] self[:] = self._gt.wait() [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] return self._exit_event.wait() [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] result = hub.switch() [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.834696] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] return self.greenlet.switch() [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] result = function(*args, **kwargs) [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] return func(*args, **kwargs) [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] raise e [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] nwinfo = self.network_api.allocate_for_instance( [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] created_port_ids = self._update_ports_for_instance( [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] with excutils.save_and_reraise_exception(): [ 628.835062] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] self.force_reraise() [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] raise self.value [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] updated_port = self._update_port( [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] _ensure_no_port_binding_failure(port) [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] raise exception.PortBindingFailed(port_id=port['id']) [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] nova.exception.PortBindingFailed: Binding failed for port 5d67aeef-bad4-4299-bc25-6652e49a217f, please check neutron logs for more information. [ 628.835514] env[62521]: ERROR nova.compute.manager [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] [ 628.835765] env[62521]: DEBUG nova.compute.utils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Binding failed for port 5d67aeef-bad4-4299-bc25-6652e49a217f, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 628.837192] env[62521]: DEBUG nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Build of instance 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4 was re-scheduled: Binding failed for port 5d67aeef-bad4-4299-bc25-6652e49a217f, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 628.837746] env[62521]: DEBUG nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 628.838080] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Acquiring lock "refresh_cache-6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.838275] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Acquired lock "refresh_cache-6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.838472] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.840040] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.137s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.844500] env[62521]: DEBUG nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 628.875920] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.876222] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.876637] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.876637] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.876831] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.877020] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.877706] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.877933] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.878159] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.878339] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.878655] env[62521]: DEBUG nova.virt.hardware [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.879398] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf0d987-e804-4792-92f1-2a360c5a96fb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.887735] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4f852a-ccdb-4bbd-b490-5859b250ed24 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.057289] env[62521]: ERROR nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e, please check neutron logs for more information. [ 629.057289] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 629.057289] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.057289] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 629.057289] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.057289] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 629.057289] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.057289] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 629.057289] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.057289] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 629.057289] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.057289] env[62521]: ERROR nova.compute.manager raise self.value [ 629.057289] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.057289] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 629.057289] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.057289] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 629.057745] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.057745] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 629.057745] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e, please check neutron logs for more information. [ 629.057745] env[62521]: ERROR nova.compute.manager [ 629.057745] env[62521]: Traceback (most recent call last): [ 629.057745] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 629.057745] env[62521]: listener.cb(fileno) [ 629.057745] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.057745] env[62521]: result = function(*args, **kwargs) [ 629.057745] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 629.057745] env[62521]: return func(*args, **kwargs) [ 629.057745] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.057745] env[62521]: raise e [ 629.057745] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.057745] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 629.057745] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.057745] env[62521]: created_port_ids = self._update_ports_for_instance( [ 629.057745] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.057745] env[62521]: with excutils.save_and_reraise_exception(): [ 629.057745] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.057745] env[62521]: self.force_reraise() [ 629.057745] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.057745] env[62521]: raise self.value [ 629.057745] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.057745] env[62521]: updated_port = self._update_port( [ 629.057745] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.057745] env[62521]: _ensure_no_port_binding_failure(port) [ 629.057745] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.057745] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 629.058410] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e, please check neutron logs for more information. [ 629.058410] env[62521]: Removing descriptor: 15 [ 629.058410] env[62521]: ERROR nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e, please check neutron logs for more information. [ 629.058410] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Traceback (most recent call last): [ 629.058410] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 629.058410] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] yield resources [ 629.058410] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.058410] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] self.driver.spawn(context, instance, image_meta, [ 629.058410] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 629.058410] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.058410] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.058410] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] vm_ref = self.build_virtual_machine(instance, [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] for vif in network_info: [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] return self._sync_wrapper(fn, *args, **kwargs) [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] self.wait() [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] self[:] = self._gt.wait() [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] return self._exit_event.wait() [ 629.058735] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] result = hub.switch() [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] return self.greenlet.switch() [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] result = function(*args, **kwargs) [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] return func(*args, **kwargs) [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] raise e [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] nwinfo = self.network_api.allocate_for_instance( [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.059084] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] created_port_ids = self._update_ports_for_instance( [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] with excutils.save_and_reraise_exception(): [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] self.force_reraise() [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] raise self.value [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] updated_port = self._update_port( [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] _ensure_no_port_binding_failure(port) [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.059449] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] raise exception.PortBindingFailed(port_id=port['id']) [ 629.059777] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] nova.exception.PortBindingFailed: Binding failed for port 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e, please check neutron logs for more information. [ 629.059777] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] [ 629.059777] env[62521]: INFO nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Terminating instance [ 629.060636] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Acquiring lock "refresh_cache-fb33026e-b607-477f-b75f-779db7c5c34f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.060636] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Acquired lock "refresh_cache-fb33026e-b607-477f-b75f-779db7c5c34f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.060755] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.090584] env[62521]: DEBUG oslo_concurrency.lockutils [req-76656e51-44cd-4f90-a3d5-21bcc8ecedc2 req-64056439-1b95-459d-a508-dd845a5dd64a service nova] Releasing lock "refresh_cache-e944e6e5-7bc0-4efb-a811-38221734ec4f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.090584] env[62521]: DEBUG nova.compute.manager [req-76656e51-44cd-4f90-a3d5-21bcc8ecedc2 req-64056439-1b95-459d-a508-dd845a5dd64a service nova] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Received event network-vif-deleted-1283c88b-3962-4229-abfd-93a41daf5b29 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.360889] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.479421] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.577077] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.635348] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.748292] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50093f33-48cc-409d-b7f2-51f93953bf87 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.756152] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545da52f-3546-45ea-82d7-34296e8bf1cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.787665] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0f0455-1b05-4c0b-a5ea-5adfb1de0ade {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.795236] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653f5f54-91fe-4e0d-bbfc-df1fc3b49127 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.808456] env[62521]: DEBUG nova.compute.provider_tree [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.983711] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Releasing lock "refresh_cache-6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.983950] env[62521]: DEBUG nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 629.984230] env[62521]: DEBUG nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.984479] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.002207] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.034736] env[62521]: DEBUG nova.compute.manager [req-5683cf32-b5f7-44b0-a2b8-6361ecc72e7a req-705624cc-2e6a-4762-8281-f7dab533f027 service nova] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Received event network-changed-1bd25481-5b47-4d60-b3b2-4d39a6d0a97e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 630.034936] env[62521]: DEBUG nova.compute.manager [req-5683cf32-b5f7-44b0-a2b8-6361ecc72e7a req-705624cc-2e6a-4762-8281-f7dab533f027 service nova] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Refreshing instance network info cache due to event network-changed-1bd25481-5b47-4d60-b3b2-4d39a6d0a97e. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 630.035141] env[62521]: DEBUG oslo_concurrency.lockutils [req-5683cf32-b5f7-44b0-a2b8-6361ecc72e7a req-705624cc-2e6a-4762-8281-f7dab533f027 service nova] Acquiring lock "refresh_cache-fb33026e-b607-477f-b75f-779db7c5c34f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.138691] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Releasing lock "refresh_cache-fb33026e-b607-477f-b75f-779db7c5c34f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.139148] env[62521]: DEBUG nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 630.139348] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 630.139662] env[62521]: DEBUG oslo_concurrency.lockutils [req-5683cf32-b5f7-44b0-a2b8-6361ecc72e7a req-705624cc-2e6a-4762-8281-f7dab533f027 service nova] Acquired lock "refresh_cache-fb33026e-b607-477f-b75f-779db7c5c34f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.139833] env[62521]: DEBUG nova.network.neutron [req-5683cf32-b5f7-44b0-a2b8-6361ecc72e7a req-705624cc-2e6a-4762-8281-f7dab533f027 service nova] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Refreshing network info cache for port 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 630.140897] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c46d16e-4cbe-440c-8ee8-89a86b4b0a3f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.152018] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6040d9-a86d-4296-aaa7-7a4f682d4722 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.175083] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fb33026e-b607-477f-b75f-779db7c5c34f could not be found. [ 630.175310] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 630.175491] env[62521]: INFO nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 630.175719] env[62521]: DEBUG oslo.service.loopingcall [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.175905] env[62521]: DEBUG nova.compute.manager [-] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.175997] env[62521]: DEBUG nova.network.neutron [-] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.189373] env[62521]: DEBUG nova.network.neutron [-] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.312037] env[62521]: DEBUG nova.scheduler.client.report [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.504543] env[62521]: DEBUG nova.network.neutron [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.662581] env[62521]: DEBUG nova.network.neutron [req-5683cf32-b5f7-44b0-a2b8-6361ecc72e7a req-705624cc-2e6a-4762-8281-f7dab533f027 service nova] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.692043] env[62521]: DEBUG nova.network.neutron [-] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.723054] env[62521]: DEBUG nova.network.neutron [req-5683cf32-b5f7-44b0-a2b8-6361ecc72e7a req-705624cc-2e6a-4762-8281-f7dab533f027 service nova] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.817162] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.977s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.817784] env[62521]: ERROR nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5e484068-db21-4dc6-9171-8e256d0e84fa, please check neutron logs for more information. [ 630.817784] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Traceback (most recent call last): [ 630.817784] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.817784] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] self.driver.spawn(context, instance, image_meta, [ 630.817784] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 630.817784] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.817784] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.817784] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] vm_ref = self.build_virtual_machine(instance, [ 630.817784] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.817784] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.817784] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] for vif in network_info: [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] return self._sync_wrapper(fn, *args, **kwargs) [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] self.wait() [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] self[:] = self._gt.wait() [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] return self._exit_event.wait() [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] current.throw(*self._exc) [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.818126] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] result = function(*args, **kwargs) [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] return func(*args, **kwargs) [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] raise e [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] nwinfo = self.network_api.allocate_for_instance( [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] created_port_ids = self._update_ports_for_instance( [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] with excutils.save_and_reraise_exception(): [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] self.force_reraise() [ 630.818453] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.818787] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] raise self.value [ 630.818787] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.818787] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] updated_port = self._update_port( [ 630.818787] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.818787] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] _ensure_no_port_binding_failure(port) [ 630.818787] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.818787] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] raise exception.PortBindingFailed(port_id=port['id']) [ 630.818787] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] nova.exception.PortBindingFailed: Binding failed for port 5e484068-db21-4dc6-9171-8e256d0e84fa, please check neutron logs for more information. [ 630.818787] env[62521]: ERROR nova.compute.manager [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] [ 630.818787] env[62521]: DEBUG nova.compute.utils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Binding failed for port 5e484068-db21-4dc6-9171-8e256d0e84fa, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 630.820140] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.753s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.820326] env[62521]: DEBUG nova.objects.instance [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62521) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 630.822860] env[62521]: DEBUG nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Build of instance 6296d4ea-d678-4981-bc5a-729e83020fa2 was re-scheduled: Binding failed for port 5e484068-db21-4dc6-9171-8e256d0e84fa, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 630.823299] env[62521]: DEBUG nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 630.823520] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquiring lock "refresh_cache-6296d4ea-d678-4981-bc5a-729e83020fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.823668] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquired lock "refresh_cache-6296d4ea-d678-4981-bc5a-729e83020fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.823825] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.007415] env[62521]: INFO nova.compute.manager [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] [instance: 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4] Took 1.02 seconds to deallocate network for instance. [ 631.194928] env[62521]: INFO nova.compute.manager [-] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Took 1.02 seconds to deallocate network for instance. [ 631.197352] env[62521]: DEBUG nova.compute.claims [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 631.197532] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.225249] env[62521]: DEBUG oslo_concurrency.lockutils [req-5683cf32-b5f7-44b0-a2b8-6361ecc72e7a req-705624cc-2e6a-4762-8281-f7dab533f027 service nova] Releasing lock "refresh_cache-fb33026e-b607-477f-b75f-779db7c5c34f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.225632] env[62521]: DEBUG nova.compute.manager [req-5683cf32-b5f7-44b0-a2b8-6361ecc72e7a req-705624cc-2e6a-4762-8281-f7dab533f027 service nova] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Received event network-vif-deleted-1bd25481-5b47-4d60-b3b2-4d39a6d0a97e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.343629] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.403246] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.831158] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e1936602-b76c-4f71-8eb3-530d78563350 tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.832349] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.046s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.834492] env[62521]: INFO nova.compute.claims [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.906062] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Releasing lock "refresh_cache-6296d4ea-d678-4981-bc5a-729e83020fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.906255] env[62521]: DEBUG nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 631.906393] env[62521]: DEBUG nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.906564] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.921548] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.039251] env[62521]: INFO nova.scheduler.client.report [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Deleted allocations for instance 6d2a2f98-aa84-4766-a872-dc6e22bcc3e4 [ 632.424379] env[62521]: DEBUG nova.network.neutron [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.548251] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7799ecbe-3dda-413d-8656-fc3e7594410d tempest-ServersWithSpecificFlavorTestJSON-571441980 tempest-ServersWithSpecificFlavorTestJSON-571441980-project-member] Lock "6d2a2f98-aa84-4766-a872-dc6e22bcc3e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.643s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.926504] env[62521]: INFO nova.compute.manager [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 6296d4ea-d678-4981-bc5a-729e83020fa2] Took 1.02 seconds to deallocate network for instance. [ 633.050404] env[62521]: DEBUG nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 633.224918] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1565f0-b7b8-40ef-b321-fb05a728c2ea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.232494] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4921f00-ad78-4c8e-beec-70f64bee331c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.262458] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6bab32e-b8a6-4543-b712-a10488f47add {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.269245] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434db78e-31fd-4075-9d3f-203ccfaf0dbc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.282183] env[62521]: DEBUG nova.compute.provider_tree [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.569422] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.785996] env[62521]: DEBUG nova.scheduler.client.report [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.959161] env[62521]: INFO nova.scheduler.client.report [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Deleted allocations for instance 6296d4ea-d678-4981-bc5a-729e83020fa2 [ 634.290226] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.290744] env[62521]: DEBUG nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 634.293517] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.794s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.467618] env[62521]: DEBUG oslo_concurrency.lockutils [None req-180a98fd-de0e-49c3-a288-1bcf459769f8 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "6296d4ea-d678-4981-bc5a-729e83020fa2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.516s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.799343] env[62521]: DEBUG nova.compute.utils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.804384] env[62521]: DEBUG nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.804622] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 634.907073] env[62521]: DEBUG nova.policy [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c926ee4cd8c4479e9fd41369634f9ead', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b64ba8aa3f124b1f84ba6e51e33538be', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.975209] env[62521]: DEBUG nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 635.251963] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Successfully created port: 876cdb7a-b71f-4b23-b194-e5cf93ec5691 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 635.278921] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d79a26-a640-4c6f-8950-f3cf2cd201a7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.286793] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330f6acb-a2ae-4373-b9a5-33ce2e3c2caa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.316114] env[62521]: DEBUG nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.319233] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b7e0d7-3fe8-4c36-b2b7-545b6c2941d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.326594] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ec5143-67d6-4259-94d2-9b82053223e6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.339424] env[62521]: DEBUG nova.compute.provider_tree [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.500449] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.843150] env[62521]: DEBUG nova.scheduler.client.report [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.184246] env[62521]: DEBUG nova.compute.manager [req-917896a7-b9b6-4300-a055-9e69601feb65 req-25895298-335d-49e8-8b47-09c71062495f service nova] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Received event network-changed-876cdb7a-b71f-4b23-b194-e5cf93ec5691 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 636.184524] env[62521]: DEBUG nova.compute.manager [req-917896a7-b9b6-4300-a055-9e69601feb65 req-25895298-335d-49e8-8b47-09c71062495f service nova] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Refreshing instance network info cache due to event network-changed-876cdb7a-b71f-4b23-b194-e5cf93ec5691. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 636.184739] env[62521]: DEBUG oslo_concurrency.lockutils [req-917896a7-b9b6-4300-a055-9e69601feb65 req-25895298-335d-49e8-8b47-09c71062495f service nova] Acquiring lock "refresh_cache-d9571cc5-0555-4867-a8d3-a2281d512305" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.184880] env[62521]: DEBUG oslo_concurrency.lockutils [req-917896a7-b9b6-4300-a055-9e69601feb65 req-25895298-335d-49e8-8b47-09c71062495f service nova] Acquired lock "refresh_cache-d9571cc5-0555-4867-a8d3-a2281d512305" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.185525] env[62521]: DEBUG nova.network.neutron [req-917896a7-b9b6-4300-a055-9e69601feb65 req-25895298-335d-49e8-8b47-09c71062495f service nova] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Refreshing network info cache for port 876cdb7a-b71f-4b23-b194-e5cf93ec5691 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 636.330476] env[62521]: DEBUG nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 636.352472] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.059s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.353314] env[62521]: ERROR nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7b451f88-f19d-4f24-9f30-a78410e4db4a, please check neutron logs for more information. [ 636.353314] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Traceback (most recent call last): [ 636.353314] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.353314] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] self.driver.spawn(context, instance, image_meta, [ 636.353314] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 636.353314] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.353314] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.353314] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] vm_ref = self.build_virtual_machine(instance, [ 636.353314] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.353314] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.353314] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] for vif in network_info: [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] return self._sync_wrapper(fn, *args, **kwargs) [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] self.wait() [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] self[:] = self._gt.wait() [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] return self._exit_event.wait() [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] result = hub.switch() [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.353788] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] return self.greenlet.switch() [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] result = function(*args, **kwargs) [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] return func(*args, **kwargs) [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] raise e [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] nwinfo = self.network_api.allocate_for_instance( [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] created_port_ids = self._update_ports_for_instance( [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] with excutils.save_and_reraise_exception(): [ 636.354404] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] self.force_reraise() [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] raise self.value [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] updated_port = self._update_port( [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] _ensure_no_port_binding_failure(port) [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] raise exception.PortBindingFailed(port_id=port['id']) [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] nova.exception.PortBindingFailed: Binding failed for port 7b451f88-f19d-4f24-9f30-a78410e4db4a, please check neutron logs for more information. [ 636.354936] env[62521]: ERROR nova.compute.manager [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] [ 636.355621] env[62521]: DEBUG nova.compute.utils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Binding failed for port 7b451f88-f19d-4f24-9f30-a78410e4db4a, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.357613] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.196s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.360967] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Build of instance 4b2fda93-a60d-4050-ae05-7a37ac284917 was re-scheduled: Binding failed for port 7b451f88-f19d-4f24-9f30-a78410e4db4a, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 636.361428] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 636.361652] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "refresh_cache-4b2fda93-a60d-4050-ae05-7a37ac284917" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.361799] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquired lock "refresh_cache-4b2fda93-a60d-4050-ae05-7a37ac284917" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.362107] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.370366] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.370592] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.370748] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.370927] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.371085] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.371236] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.371441] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.371596] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.371759] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.371919] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.372108] env[62521]: DEBUG nova.virt.hardware [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.373171] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fac0a4-e148-48cc-85ae-529212a9d4f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.386093] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bddbf6ca-ab54-4d52-a7e8-d68efd2e21d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.401467] env[62521]: ERROR nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 876cdb7a-b71f-4b23-b194-e5cf93ec5691, please check neutron logs for more information. [ 636.401467] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 636.401467] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.401467] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 636.401467] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.401467] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 636.401467] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.401467] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 636.401467] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.401467] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 636.401467] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.401467] env[62521]: ERROR nova.compute.manager raise self.value [ 636.401467] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.401467] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 636.401467] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.401467] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 636.401906] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.401906] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 636.401906] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 876cdb7a-b71f-4b23-b194-e5cf93ec5691, please check neutron logs for more information. [ 636.401906] env[62521]: ERROR nova.compute.manager [ 636.401906] env[62521]: Traceback (most recent call last): [ 636.401906] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 636.401906] env[62521]: listener.cb(fileno) [ 636.401906] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.401906] env[62521]: result = function(*args, **kwargs) [ 636.401906] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 636.401906] env[62521]: return func(*args, **kwargs) [ 636.401906] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.401906] env[62521]: raise e [ 636.401906] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.401906] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 636.401906] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.401906] env[62521]: created_port_ids = self._update_ports_for_instance( [ 636.401906] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.401906] env[62521]: with excutils.save_and_reraise_exception(): [ 636.401906] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.401906] env[62521]: self.force_reraise() [ 636.401906] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.401906] env[62521]: raise self.value [ 636.401906] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.401906] env[62521]: updated_port = self._update_port( [ 636.401906] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.401906] env[62521]: _ensure_no_port_binding_failure(port) [ 636.401906] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.401906] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 636.402661] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 876cdb7a-b71f-4b23-b194-e5cf93ec5691, please check neutron logs for more information. [ 636.402661] env[62521]: Removing descriptor: 15 [ 636.402661] env[62521]: ERROR nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 876cdb7a-b71f-4b23-b194-e5cf93ec5691, please check neutron logs for more information. [ 636.402661] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Traceback (most recent call last): [ 636.402661] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 636.402661] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] yield resources [ 636.402661] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.402661] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] self.driver.spawn(context, instance, image_meta, [ 636.402661] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 636.402661] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.402661] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.402661] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] vm_ref = self.build_virtual_machine(instance, [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] for vif in network_info: [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] return self._sync_wrapper(fn, *args, **kwargs) [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] self.wait() [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] self[:] = self._gt.wait() [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] return self._exit_event.wait() [ 636.402964] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] result = hub.switch() [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] return self.greenlet.switch() [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] result = function(*args, **kwargs) [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] return func(*args, **kwargs) [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] raise e [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] nwinfo = self.network_api.allocate_for_instance( [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.403315] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] created_port_ids = self._update_ports_for_instance( [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] with excutils.save_and_reraise_exception(): [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] self.force_reraise() [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] raise self.value [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] updated_port = self._update_port( [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] _ensure_no_port_binding_failure(port) [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.403631] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] raise exception.PortBindingFailed(port_id=port['id']) [ 636.403921] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] nova.exception.PortBindingFailed: Binding failed for port 876cdb7a-b71f-4b23-b194-e5cf93ec5691, please check neutron logs for more information. [ 636.403921] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] [ 636.403921] env[62521]: INFO nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Terminating instance [ 636.404655] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquiring lock "refresh_cache-d9571cc5-0555-4867-a8d3-a2281d512305" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.710498] env[62521]: DEBUG nova.network.neutron [req-917896a7-b9b6-4300-a055-9e69601feb65 req-25895298-335d-49e8-8b47-09c71062495f service nova] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.811711] env[62521]: DEBUG nova.network.neutron [req-917896a7-b9b6-4300-a055-9e69601feb65 req-25895298-335d-49e8-8b47-09c71062495f service nova] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.889731] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.015964] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.303842] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82ca7f7-6eb9-4fc7-b976-162567ebc9e4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.879340] env[62521]: DEBUG oslo_concurrency.lockutils [req-917896a7-b9b6-4300-a055-9e69601feb65 req-25895298-335d-49e8-8b47-09c71062495f service nova] Releasing lock "refresh_cache-d9571cc5-0555-4867-a8d3-a2281d512305" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.879755] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Releasing lock "refresh_cache-4b2fda93-a60d-4050-ae05-7a37ac284917" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.880021] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 637.880126] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.880287] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.882351] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquired lock "refresh_cache-d9571cc5-0555-4867-a8d3-a2281d512305" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.882830] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.884266] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5e0443-cdcc-4125-86d3-c05aeb1d4478 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.915180] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.916790] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a083a1e2-5345-40ff-aa33-bc945742c546 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.925063] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea78b65-8bd3-42cd-b5f6-012cb96889fe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.938560] env[62521]: DEBUG nova.compute.provider_tree [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.267512] env[62521]: DEBUG nova.compute.manager [req-d2f97c3f-aa2a-4415-b402-271e8db342db req-2d1af064-c9be-4339-a6d5-10eca8468ff7 service nova] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Received event network-vif-deleted-876cdb7a-b71f-4b23-b194-e5cf93ec5691 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 638.420150] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.445960] env[62521]: DEBUG nova.scheduler.client.report [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.451078] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.562926] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.923779] env[62521]: INFO nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 4b2fda93-a60d-4050-ae05-7a37ac284917] Took 1.04 seconds to deallocate network for instance. [ 638.954749] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.599s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.955449] env[62521]: ERROR nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4ae88467-3176-442d-a60a-297ed80fae99, please check neutron logs for more information. [ 638.955449] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Traceback (most recent call last): [ 638.955449] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.955449] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] self.driver.spawn(context, instance, image_meta, [ 638.955449] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 638.955449] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.955449] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.955449] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] vm_ref = self.build_virtual_machine(instance, [ 638.955449] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.955449] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.955449] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] for vif in network_info: [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] return self._sync_wrapper(fn, *args, **kwargs) [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] self.wait() [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] self[:] = self._gt.wait() [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] return self._exit_event.wait() [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] result = hub.switch() [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.955723] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] return self.greenlet.switch() [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] result = function(*args, **kwargs) [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] return func(*args, **kwargs) [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] raise e [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] nwinfo = self.network_api.allocate_for_instance( [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] created_port_ids = self._update_ports_for_instance( [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] with excutils.save_and_reraise_exception(): [ 638.956077] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] self.force_reraise() [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] raise self.value [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] updated_port = self._update_port( [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] _ensure_no_port_binding_failure(port) [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] raise exception.PortBindingFailed(port_id=port['id']) [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] nova.exception.PortBindingFailed: Binding failed for port 4ae88467-3176-442d-a60a-297ed80fae99, please check neutron logs for more information. [ 638.956427] env[62521]: ERROR nova.compute.manager [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] [ 638.956685] env[62521]: DEBUG nova.compute.utils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Binding failed for port 4ae88467-3176-442d-a60a-297ed80fae99, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.957375] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.927s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.957656] env[62521]: DEBUG nova.objects.instance [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lazy-loading 'resources' on Instance uuid 20b01df0-4ced-4668-a786-046251a0c7ed {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 638.959029] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Build of instance 1d25cdca-3c87-47a2-85aa-86f7132297b5 was re-scheduled: Binding failed for port 4ae88467-3176-442d-a60a-297ed80fae99, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 638.959550] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 638.959809] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "refresh_cache-1d25cdca-3c87-47a2-85aa-86f7132297b5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.959973] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquired lock "refresh_cache-1d25cdca-3c87-47a2-85aa-86f7132297b5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.960156] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 639.065935] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Releasing lock "refresh_cache-d9571cc5-0555-4867-a8d3-a2281d512305" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.066400] env[62521]: DEBUG nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 639.066596] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.066900] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c3212577-8e9a-47fd-bd42-961665fa96b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.075683] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27ca00b-113f-4d90-85af-3bcaf1384e71 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.097130] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d9571cc5-0555-4867-a8d3-a2281d512305 could not be found. [ 639.097369] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.097553] env[62521]: INFO nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Took 0.03 seconds to destroy the instance on the hypervisor. [ 639.097789] env[62521]: DEBUG oslo.service.loopingcall [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.098347] env[62521]: DEBUG nova.compute.manager [-] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.098447] env[62521]: DEBUG nova.network.neutron [-] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.120776] env[62521]: DEBUG nova.network.neutron [-] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.481716] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.592973] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.626226] env[62521]: DEBUG nova.network.neutron [-] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.888410] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dcc2de3-433b-4600-8eb1-44f685257c67 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.896371] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b68e24c-2198-4a98-9021-c80368a17a64 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.927710] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ef032d-311e-4f10-84bd-8725f2d8d120 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.938092] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abd63d4-3917-4618-b6ea-4fed9676537a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.952951] env[62521]: DEBUG nova.compute.provider_tree [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.961727] env[62521]: INFO nova.scheduler.client.report [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Deleted allocations for instance 4b2fda93-a60d-4050-ae05-7a37ac284917 [ 640.096110] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Releasing lock "refresh_cache-1d25cdca-3c87-47a2-85aa-86f7132297b5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.096380] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 640.096607] env[62521]: DEBUG nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.096736] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 640.128447] env[62521]: INFO nova.compute.manager [-] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Took 1.03 seconds to deallocate network for instance. [ 640.130738] env[62521]: DEBUG nova.compute.claims [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.130913] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.137906] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.456643] env[62521]: DEBUG nova.scheduler.client.report [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.469293] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "4b2fda93-a60d-4050-ae05-7a37ac284917" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.832s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.640416] env[62521]: DEBUG nova.network.neutron [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.961758] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.004s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.964017] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.569s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.971854] env[62521]: DEBUG nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 640.994034] env[62521]: INFO nova.scheduler.client.report [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Deleted allocations for instance 20b01df0-4ced-4668-a786-046251a0c7ed [ 641.143222] env[62521]: INFO nova.compute.manager [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 1d25cdca-3c87-47a2-85aa-86f7132297b5] Took 1.05 seconds to deallocate network for instance. [ 641.490169] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.501530] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5172a7fc-2f54-4af6-bcbb-fa85488c4f5f tempest-ServerShowV257Test-1594865257 tempest-ServerShowV257Test-1594865257-project-member] Lock "20b01df0-4ced-4668-a786-046251a0c7ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.232s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.884142] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42f308d-7890-4916-9451-e855511be7d1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.892101] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a8519d-dad8-4f4c-96d6-499aaef1a89b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.924825] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714664b2-a1e4-46e0-a023-373f9ecbf07b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.933953] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9693861-c3b0-4c14-b16e-2b3240009788 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.949163] env[62521]: DEBUG nova.compute.provider_tree [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.177490] env[62521]: INFO nova.scheduler.client.report [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Deleted allocations for instance 1d25cdca-3c87-47a2-85aa-86f7132297b5 [ 642.451020] env[62521]: DEBUG nova.scheduler.client.report [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.686099] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bba3c794-b344-4c07-98be-e7ce44c85dde tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "1d25cdca-3c87-47a2-85aa-86f7132297b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.009s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.960837] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.960837] env[62521]: ERROR nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d218a569-f06e-4f89-934c-288f23df0e46, please check neutron logs for more information. [ 642.960837] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] Traceback (most recent call last): [ 642.960837] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.960837] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] self.driver.spawn(context, instance, image_meta, [ 642.960837] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 642.960837] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.960837] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.960837] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] vm_ref = self.build_virtual_machine(instance, [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] for vif in network_info: [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] return self._sync_wrapper(fn, *args, **kwargs) [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] self.wait() [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] self[:] = self._gt.wait() [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] return self._exit_event.wait() [ 642.961880] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] current.throw(*self._exc) [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] result = function(*args, **kwargs) [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] return func(*args, **kwargs) [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] raise e [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] nwinfo = self.network_api.allocate_for_instance( [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] created_port_ids = self._update_ports_for_instance( [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.962216] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] with excutils.save_and_reraise_exception(): [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] self.force_reraise() [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] raise self.value [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] updated_port = self._update_port( [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] _ensure_no_port_binding_failure(port) [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] raise exception.PortBindingFailed(port_id=port['id']) [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] nova.exception.PortBindingFailed: Binding failed for port d218a569-f06e-4f89-934c-288f23df0e46, please check neutron logs for more information. [ 642.962520] env[62521]: ERROR nova.compute.manager [instance: 1c892e78-a863-4561-940d-b44701e97041] [ 642.962806] env[62521]: DEBUG nova.compute.utils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Binding failed for port d218a569-f06e-4f89-934c-288f23df0e46, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 642.962806] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.476s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.967109] env[62521]: DEBUG nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Build of instance 1c892e78-a863-4561-940d-b44701e97041 was re-scheduled: Binding failed for port d218a569-f06e-4f89-934c-288f23df0e46, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 642.967109] env[62521]: DEBUG nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 642.967109] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquiring lock "refresh_cache-1c892e78-a863-4561-940d-b44701e97041" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.967303] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Acquired lock "refresh_cache-1c892e78-a863-4561-940d-b44701e97041" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.968737] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.188468] env[62521]: DEBUG nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 643.344973] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Acquiring lock "f6a27e36-1728-4056-99a6-ed2114df64ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.345185] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Lock "f6a27e36-1728-4056-99a6-ed2114df64ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.488421] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.567111] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.712321] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.865053] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca49d81-bc37-40e4-8472-1b855c795bd9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.872893] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3278cdc8-2561-4f73-b190-fc088b573040 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.902756] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f05cae2-fc94-43f0-9816-047a83944b0f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.910440] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbeb28b9-d83b-418f-9291-abf86891f71e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.924702] env[62521]: DEBUG nova.compute.provider_tree [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.069110] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Releasing lock "refresh_cache-1c892e78-a863-4561-940d-b44701e97041" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.069383] env[62521]: DEBUG nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 644.069553] env[62521]: DEBUG nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.070253] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 644.095345] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.223068] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 644.223337] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 644.428828] env[62521]: DEBUG nova.scheduler.client.report [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.516502] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "6be7d826-e19b-431e-b714-25e646e6b7ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.516735] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "6be7d826-e19b-431e-b714-25e646e6b7ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.541749] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "590b2441-e627-463b-9722-d0fcb8fcd83a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.541965] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "590b2441-e627-463b-9722-d0fcb8fcd83a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.597912] env[62521]: DEBUG nova.network.neutron [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.728759] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 644.728997] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 644.729137] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Rebuilding the list of instances to heal {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 644.935840] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.936497] env[62521]: ERROR nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 239afdcf-1140-44fd-9280-cfb1580f845c, please check neutron logs for more information. [ 644.936497] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Traceback (most recent call last): [ 644.936497] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.936497] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] self.driver.spawn(context, instance, image_meta, [ 644.936497] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 644.936497] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.936497] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.936497] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] vm_ref = self.build_virtual_machine(instance, [ 644.936497] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.936497] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.936497] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] for vif in network_info: [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] return self._sync_wrapper(fn, *args, **kwargs) [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] self.wait() [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] self[:] = self._gt.wait() [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] return self._exit_event.wait() [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] current.throw(*self._exc) [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.936826] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] result = function(*args, **kwargs) [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] return func(*args, **kwargs) [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] raise e [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] nwinfo = self.network_api.allocate_for_instance( [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] created_port_ids = self._update_ports_for_instance( [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] with excutils.save_and_reraise_exception(): [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] self.force_reraise() [ 644.937199] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.937599] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] raise self.value [ 644.937599] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.937599] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] updated_port = self._update_port( [ 644.937599] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.937599] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] _ensure_no_port_binding_failure(port) [ 644.937599] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.937599] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] raise exception.PortBindingFailed(port_id=port['id']) [ 644.937599] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] nova.exception.PortBindingFailed: Binding failed for port 239afdcf-1140-44fd-9280-cfb1580f845c, please check neutron logs for more information. [ 644.937599] env[62521]: ERROR nova.compute.manager [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] [ 644.937599] env[62521]: DEBUG nova.compute.utils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Binding failed for port 239afdcf-1140-44fd-9280-cfb1580f845c, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 644.938467] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.031s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.938953] env[62521]: DEBUG nova.objects.instance [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lazy-loading 'resources' on Instance uuid 7653f469-5ce5-4d74-9a3f-55a6adb3a255 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 644.940200] env[62521]: DEBUG nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Build of instance 9c2c0e66-5c67-4e0a-9205-4080bda44a66 was re-scheduled: Binding failed for port 239afdcf-1140-44fd-9280-cfb1580f845c, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 644.940613] env[62521]: DEBUG nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 644.940843] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Acquiring lock "refresh_cache-9c2c0e66-5c67-4e0a-9205-4080bda44a66" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.940987] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Acquired lock "refresh_cache-9c2c0e66-5c67-4e0a-9205-4080bda44a66" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.941157] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.101038] env[62521]: INFO nova.compute.manager [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] [instance: 1c892e78-a863-4561-940d-b44701e97041] Took 1.03 seconds to deallocate network for instance. [ 645.234889] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 645.234889] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 645.234889] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 645.234889] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 645.253492] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "refresh_cache-7653f469-5ce5-4d74-9a3f-55a6adb3a255" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.253643] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquired lock "refresh_cache-7653f469-5ce5-4d74-9a3f-55a6adb3a255" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.253791] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Forcefully refreshing network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 645.253950] env[62521]: DEBUG nova.objects.instance [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lazy-loading 'info_cache' on Instance uuid 7653f469-5ce5-4d74-9a3f-55a6adb3a255 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 645.461723] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.556149] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.808718] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5f6b95-6a62-472d-a369-9f3d8257fce7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.817124] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98dd3cfe-1e2c-48ed-a0ef-71cf9ff314f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.846822] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff1450a-99cc-49a9-8ae9-9d4e142df489 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.854008] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f27e5d-df73-48a7-a52b-916fb77ae323 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.866965] env[62521]: DEBUG nova.compute.provider_tree [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.057589] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Releasing lock "refresh_cache-9c2c0e66-5c67-4e0a-9205-4080bda44a66" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.057804] env[62521]: DEBUG nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 646.057988] env[62521]: DEBUG nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.058176] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.076697] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.129598] env[62521]: INFO nova.scheduler.client.report [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Deleted allocations for instance 1c892e78-a863-4561-940d-b44701e97041 [ 646.277174] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.370049] env[62521]: DEBUG nova.scheduler.client.report [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.578965] env[62521]: DEBUG nova.network.neutron [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.640608] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9da6ee5b-d103-4982-9f7f-8de7c1f77b50 tempest-ListImageFiltersTestJSON-2037507189 tempest-ListImageFiltersTestJSON-2037507189-project-member] Lock "1c892e78-a863-4561-940d-b44701e97041" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.259s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.850317] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.875737] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.937s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.877573] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.736s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.904997] env[62521]: INFO nova.scheduler.client.report [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Deleted allocations for instance 7653f469-5ce5-4d74-9a3f-55a6adb3a255 [ 647.082213] env[62521]: INFO nova.compute.manager [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] [instance: 9c2c0e66-5c67-4e0a-9205-4080bda44a66] Took 1.02 seconds to deallocate network for instance. [ 647.142974] env[62521]: DEBUG nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 647.352467] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Releasing lock "refresh_cache-7653f469-5ce5-4d74-9a3f-55a6adb3a255" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.352735] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Updated the network info_cache for instance {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 647.352910] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.353086] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.353239] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.353389] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.353533] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.353679] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.353806] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 647.353944] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.413752] env[62521]: DEBUG oslo_concurrency.lockutils [None req-53877a0d-aea3-4426-9570-16ed15d172c9 tempest-ServerDiagnosticsV248Test-2117754024 tempest-ServerDiagnosticsV248Test-2117754024-project-member] Lock "7653f469-5ce5-4d74-9a3f-55a6adb3a255" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.250s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.665509] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.780440] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a525f7-3f56-4e1b-9b1e-af8e4050df93 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.788966] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f48078-a23c-461f-8c0d-185f72af23f0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.821456] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeeacd65-ce3c-4d54-80aa-a40ea72ab460 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.828688] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251c4e59-a1cb-417a-b1b5-88c7b9f9157c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.841606] env[62521]: DEBUG nova.compute.provider_tree [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.857055] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.116048] env[62521]: INFO nova.scheduler.client.report [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Deleted allocations for instance 9c2c0e66-5c67-4e0a-9205-4080bda44a66 [ 648.344895] env[62521]: DEBUG nova.scheduler.client.report [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.625712] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b01881-5065-416d-bea9-e24d1841f6b2 tempest-AttachInterfacesV270Test-1178669164 tempest-AttachInterfacesV270Test-1178669164-project-member] Lock "9c2c0e66-5c67-4e0a-9205-4080bda44a66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.164s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.850546] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.973s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.851484] env[62521]: ERROR nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f30572e-7dc3-4725-943c-feab59bb48ef, please check neutron logs for more information. [ 648.851484] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Traceback (most recent call last): [ 648.851484] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.851484] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] self.driver.spawn(context, instance, image_meta, [ 648.851484] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 648.851484] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.851484] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.851484] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] vm_ref = self.build_virtual_machine(instance, [ 648.851484] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.851484] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.851484] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] for vif in network_info: [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] return self._sync_wrapper(fn, *args, **kwargs) [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] self.wait() [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] self[:] = self._gt.wait() [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] return self._exit_event.wait() [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] result = hub.switch() [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.851819] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] return self.greenlet.switch() [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] result = function(*args, **kwargs) [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] return func(*args, **kwargs) [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] raise e [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] nwinfo = self.network_api.allocate_for_instance( [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] created_port_ids = self._update_ports_for_instance( [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] with excutils.save_and_reraise_exception(): [ 648.852136] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] self.force_reraise() [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] raise self.value [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] updated_port = self._update_port( [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] _ensure_no_port_binding_failure(port) [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] raise exception.PortBindingFailed(port_id=port['id']) [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] nova.exception.PortBindingFailed: Binding failed for port 1f30572e-7dc3-4725-943c-feab59bb48ef, please check neutron logs for more information. [ 648.852662] env[62521]: ERROR nova.compute.manager [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] [ 648.852921] env[62521]: DEBUG nova.compute.utils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Binding failed for port 1f30572e-7dc3-4725-943c-feab59bb48ef, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 648.853393] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.244s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.856955] env[62521]: DEBUG nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Build of instance da9d2279-79ae-4371-bef5-08ae15f2d9c9 was re-scheduled: Binding failed for port 1f30572e-7dc3-4725-943c-feab59bb48ef, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 648.859012] env[62521]: DEBUG nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 648.859012] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Acquiring lock "refresh_cache-da9d2279-79ae-4371-bef5-08ae15f2d9c9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.859012] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Acquired lock "refresh_cache-da9d2279-79ae-4371-bef5-08ae15f2d9c9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.859012] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.129397] env[62521]: DEBUG nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 649.620742] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.649151] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.803722] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.893220] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a676d31d-121c-4420-8423-2c97c354ee75 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.902095] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27cd40cc-ea1f-42bb-98af-14da6faad972 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.935150] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83dbdf03-5155-4937-8ef3-2a8bfe8be3c4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.948076] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6dc0bd3-8e7d-410d-a5fd-33585fdd64b1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.962173] env[62521]: DEBUG nova.compute.provider_tree [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.309286] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Releasing lock "refresh_cache-da9d2279-79ae-4371-bef5-08ae15f2d9c9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.309286] env[62521]: DEBUG nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 650.309286] env[62521]: DEBUG nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.309286] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.340293] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.466038] env[62521]: DEBUG nova.scheduler.client.report [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.844291] env[62521]: DEBUG nova.network.neutron [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.971191] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.118s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.971830] env[62521]: ERROR nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1283c88b-3962-4229-abfd-93a41daf5b29, please check neutron logs for more information. [ 650.971830] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Traceback (most recent call last): [ 650.971830] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.971830] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] self.driver.spawn(context, instance, image_meta, [ 650.971830] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 650.971830] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.971830] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.971830] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] vm_ref = self.build_virtual_machine(instance, [ 650.971830] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.971830] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.971830] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] for vif in network_info: [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] return self._sync_wrapper(fn, *args, **kwargs) [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] self.wait() [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] self[:] = self._gt.wait() [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] return self._exit_event.wait() [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] result = hub.switch() [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.972331] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] return self.greenlet.switch() [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] result = function(*args, **kwargs) [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] return func(*args, **kwargs) [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] raise e [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] nwinfo = self.network_api.allocate_for_instance( [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] created_port_ids = self._update_ports_for_instance( [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] with excutils.save_and_reraise_exception(): [ 650.972878] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] self.force_reraise() [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] raise self.value [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] updated_port = self._update_port( [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] _ensure_no_port_binding_failure(port) [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] raise exception.PortBindingFailed(port_id=port['id']) [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] nova.exception.PortBindingFailed: Binding failed for port 1283c88b-3962-4229-abfd-93a41daf5b29, please check neutron logs for more information. [ 650.973410] env[62521]: ERROR nova.compute.manager [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] [ 650.973854] env[62521]: DEBUG nova.compute.utils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Binding failed for port 1283c88b-3962-4229-abfd-93a41daf5b29, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 650.974213] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.777s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.977262] env[62521]: DEBUG nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Build of instance e944e6e5-7bc0-4efb-a811-38221734ec4f was re-scheduled: Binding failed for port 1283c88b-3962-4229-abfd-93a41daf5b29, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 650.977708] env[62521]: DEBUG nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 650.977929] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Acquiring lock "refresh_cache-e944e6e5-7bc0-4efb-a811-38221734ec4f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.978086] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Acquired lock "refresh_cache-e944e6e5-7bc0-4efb-a811-38221734ec4f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.978422] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.351275] env[62521]: INFO nova.compute.manager [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] [instance: da9d2279-79ae-4371-bef5-08ae15f2d9c9] Took 1.04 seconds to deallocate network for instance. [ 651.506577] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.621254] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.884160] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c1a728-cab8-4473-8186-b3df29ba3994 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.893308] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcaf94a2-06d4-4cce-89aa-ff4c43358f75 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.927588] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4028b69c-f464-49bf-822f-ae16a31afc4a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.937228] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a19ff24-371c-490d-812f-d84e5be656fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.952868] env[62521]: DEBUG nova.compute.provider_tree [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.130745] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Releasing lock "refresh_cache-e944e6e5-7bc0-4efb-a811-38221734ec4f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.130745] env[62521]: DEBUG nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 652.130745] env[62521]: DEBUG nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.130745] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.146533] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.395511] env[62521]: INFO nova.scheduler.client.report [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Deleted allocations for instance da9d2279-79ae-4371-bef5-08ae15f2d9c9 [ 652.457998] env[62521]: DEBUG nova.scheduler.client.report [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.609715] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "062f311c-a783-41cf-8d83-b3a8d4df14f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.609966] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "062f311c-a783-41cf-8d83-b3a8d4df14f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.650913] env[62521]: DEBUG nova.network.neutron [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.907384] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a013058-ff4b-4d23-821d-bfe8753c2598 tempest-ServerActionsTestOtherA-525285490 tempest-ServerActionsTestOtherA-525285490-project-member] Lock "da9d2279-79ae-4371-bef5-08ae15f2d9c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.342s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.963794] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.989s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.964549] env[62521]: ERROR nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e, please check neutron logs for more information. [ 652.964549] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Traceback (most recent call last): [ 652.964549] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.964549] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] self.driver.spawn(context, instance, image_meta, [ 652.964549] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 652.964549] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.964549] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.964549] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] vm_ref = self.build_virtual_machine(instance, [ 652.964549] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.964549] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.964549] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] for vif in network_info: [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] return self._sync_wrapper(fn, *args, **kwargs) [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] self.wait() [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] self[:] = self._gt.wait() [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] return self._exit_event.wait() [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] result = hub.switch() [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.966412] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] return self.greenlet.switch() [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] result = function(*args, **kwargs) [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] return func(*args, **kwargs) [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] raise e [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] nwinfo = self.network_api.allocate_for_instance( [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] created_port_ids = self._update_ports_for_instance( [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] with excutils.save_and_reraise_exception(): [ 652.966768] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] self.force_reraise() [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] raise self.value [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] updated_port = self._update_port( [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] _ensure_no_port_binding_failure(port) [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] raise exception.PortBindingFailed(port_id=port['id']) [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] nova.exception.PortBindingFailed: Binding failed for port 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e, please check neutron logs for more information. [ 652.967151] env[62521]: ERROR nova.compute.manager [instance: fb33026e-b607-477f-b75f-779db7c5c34f] [ 652.967418] env[62521]: DEBUG nova.compute.utils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Binding failed for port 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 652.967418] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.398s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.969036] env[62521]: INFO nova.compute.claims [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.971914] env[62521]: DEBUG nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Build of instance fb33026e-b607-477f-b75f-779db7c5c34f was re-scheduled: Binding failed for port 1bd25481-5b47-4d60-b3b2-4d39a6d0a97e, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 652.972453] env[62521]: DEBUG nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 652.973578] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Acquiring lock "refresh_cache-fb33026e-b607-477f-b75f-779db7c5c34f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.973578] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Acquired lock "refresh_cache-fb33026e-b607-477f-b75f-779db7c5c34f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.973578] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.012370] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "868c7798-8d0f-4c31-88dc-5007454796f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.012719] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "868c7798-8d0f-4c31-88dc-5007454796f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.156176] env[62521]: INFO nova.compute.manager [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] [instance: e944e6e5-7bc0-4efb-a811-38221734ec4f] Took 1.03 seconds to deallocate network for instance. [ 653.411567] env[62521]: DEBUG nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 653.521222] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.635376] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.955774] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.143547] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Releasing lock "refresh_cache-fb33026e-b607-477f-b75f-779db7c5c34f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.143865] env[62521]: DEBUG nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 654.144103] env[62521]: DEBUG nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.144325] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 654.196917] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.201473] env[62521]: INFO nova.scheduler.client.report [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Deleted allocations for instance e944e6e5-7bc0-4efb-a811-38221734ec4f [ 654.533417] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa5be9d-57af-4762-8594-44a8fdd08f9d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.543161] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15dac25-bf22-4221-9b79-d62c4a99c3aa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.575583] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7821977d-0a71-4f3e-858c-85e21fa66be2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.584494] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc36d2d-f37f-4207-a4b5-2d6ca1f2f054 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.597053] env[62521]: DEBUG nova.compute.provider_tree [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.711431] env[62521]: DEBUG nova.network.neutron [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.714204] env[62521]: DEBUG oslo_concurrency.lockutils [None req-660f9e1d-4b39-407e-8755-586d4b078ef6 tempest-ServersTestManualDisk-2132009043 tempest-ServersTestManualDisk-2132009043-project-member] Lock "e944e6e5-7bc0-4efb-a811-38221734ec4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.026s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.101096] env[62521]: DEBUG nova.scheduler.client.report [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.219829] env[62521]: INFO nova.compute.manager [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] [instance: fb33026e-b607-477f-b75f-779db7c5c34f] Took 1.07 seconds to deallocate network for instance. [ 655.224162] env[62521]: DEBUG nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 655.608435] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.641s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.610614] env[62521]: DEBUG nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 655.613810] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.114s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.615452] env[62521]: INFO nova.compute.claims [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.770528] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.127478] env[62521]: DEBUG nova.compute.utils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 656.127478] env[62521]: DEBUG nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 656.127478] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 656.245600] env[62521]: DEBUG nova.policy [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c926ee4cd8c4479e9fd41369634f9ead', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b64ba8aa3f124b1f84ba6e51e33538be', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 656.265028] env[62521]: INFO nova.scheduler.client.report [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Deleted allocations for instance fb33026e-b607-477f-b75f-779db7c5c34f [ 656.633705] env[62521]: DEBUG nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 656.674549] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Successfully created port: c84fc1ca-c5e3-40ba-a1d2-265501cf1467 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 656.775707] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae465147-38d1-465c-a720-f6d505a90c86 tempest-ServerDiagnosticsNegativeTest-6956461 tempest-ServerDiagnosticsNegativeTest-6956461-project-member] Lock "fb33026e-b607-477f-b75f-779db7c5c34f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.688s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.148597] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Acquiring lock "de583c0b-d147-4f54-bc90-49bdd71c56aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.149228] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Lock "de583c0b-d147-4f54-bc90-49bdd71c56aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.172686] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea837b7-fe13-4d8e-acbb-1dd57101bc3a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.180484] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0650f75-0bb6-45f8-b2a1-2c6f82ac353c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.213832] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fa573d-cb8a-4fb0-8b2e-aaf26dd26833 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.223117] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1714b27-82ca-41bf-a290-6634391a0140 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.238011] env[62521]: DEBUG nova.compute.provider_tree [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.278766] env[62521]: DEBUG nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 657.651999] env[62521]: DEBUG nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 657.688846] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.689115] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.689273] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.689452] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.689599] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.689741] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.689945] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.690222] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.690419] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.690578] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.690825] env[62521]: DEBUG nova.virt.hardware [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.691711] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53912c47-c483-4535-872d-917bc03436a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.699631] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd293106-953d-4c0d-ae76-eb5b5f1006ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.741474] env[62521]: DEBUG nova.scheduler.client.report [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.804804] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.208580] env[62521]: DEBUG nova.compute.manager [req-b51c23d5-203b-45e3-89f6-2524d863ab48 req-2091e028-ad79-494e-b6b7-6f0692a1d675 service nova] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Received event network-changed-c84fc1ca-c5e3-40ba-a1d2-265501cf1467 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.208852] env[62521]: DEBUG nova.compute.manager [req-b51c23d5-203b-45e3-89f6-2524d863ab48 req-2091e028-ad79-494e-b6b7-6f0692a1d675 service nova] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Refreshing instance network info cache due to event network-changed-c84fc1ca-c5e3-40ba-a1d2-265501cf1467. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 658.208931] env[62521]: DEBUG oslo_concurrency.lockutils [req-b51c23d5-203b-45e3-89f6-2524d863ab48 req-2091e028-ad79-494e-b6b7-6f0692a1d675 service nova] Acquiring lock "refresh_cache-fe69caab-dd71-4b01-a121-2591111b193d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.209216] env[62521]: DEBUG oslo_concurrency.lockutils [req-b51c23d5-203b-45e3-89f6-2524d863ab48 req-2091e028-ad79-494e-b6b7-6f0692a1d675 service nova] Acquired lock "refresh_cache-fe69caab-dd71-4b01-a121-2591111b193d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.209254] env[62521]: DEBUG nova.network.neutron [req-b51c23d5-203b-45e3-89f6-2524d863ab48 req-2091e028-ad79-494e-b6b7-6f0692a1d675 service nova] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Refreshing network info cache for port c84fc1ca-c5e3-40ba-a1d2-265501cf1467 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 658.248677] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.634s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.249861] env[62521]: DEBUG nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 658.252069] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.121s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.492669] env[62521]: ERROR nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c84fc1ca-c5e3-40ba-a1d2-265501cf1467, please check neutron logs for more information. [ 658.492669] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 658.492669] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.492669] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 658.492669] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.492669] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 658.492669] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.492669] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 658.492669] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.492669] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 658.492669] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.492669] env[62521]: ERROR nova.compute.manager raise self.value [ 658.492669] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.492669] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 658.492669] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.492669] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 658.493200] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.493200] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 658.493200] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c84fc1ca-c5e3-40ba-a1d2-265501cf1467, please check neutron logs for more information. [ 658.493200] env[62521]: ERROR nova.compute.manager [ 658.493200] env[62521]: Traceback (most recent call last): [ 658.493200] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 658.493200] env[62521]: listener.cb(fileno) [ 658.493200] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.493200] env[62521]: result = function(*args, **kwargs) [ 658.493200] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.493200] env[62521]: return func(*args, **kwargs) [ 658.493200] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.493200] env[62521]: raise e [ 658.493200] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.493200] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 658.493200] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.493200] env[62521]: created_port_ids = self._update_ports_for_instance( [ 658.493200] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.493200] env[62521]: with excutils.save_and_reraise_exception(): [ 658.493200] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.493200] env[62521]: self.force_reraise() [ 658.493200] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.493200] env[62521]: raise self.value [ 658.493200] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.493200] env[62521]: updated_port = self._update_port( [ 658.493200] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.493200] env[62521]: _ensure_no_port_binding_failure(port) [ 658.493200] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.493200] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 658.493915] env[62521]: nova.exception.PortBindingFailed: Binding failed for port c84fc1ca-c5e3-40ba-a1d2-265501cf1467, please check neutron logs for more information. [ 658.493915] env[62521]: Removing descriptor: 15 [ 658.493915] env[62521]: ERROR nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c84fc1ca-c5e3-40ba-a1d2-265501cf1467, please check neutron logs for more information. [ 658.493915] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] Traceback (most recent call last): [ 658.493915] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 658.493915] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] yield resources [ 658.493915] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 658.493915] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] self.driver.spawn(context, instance, image_meta, [ 658.493915] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 658.493915] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.493915] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.493915] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] vm_ref = self.build_virtual_machine(instance, [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] for vif in network_info: [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] return self._sync_wrapper(fn, *args, **kwargs) [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] self.wait() [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] self[:] = self._gt.wait() [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] return self._exit_event.wait() [ 658.494210] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] result = hub.switch() [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] return self.greenlet.switch() [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] result = function(*args, **kwargs) [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] return func(*args, **kwargs) [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] raise e [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] nwinfo = self.network_api.allocate_for_instance( [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.494964] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] created_port_ids = self._update_ports_for_instance( [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] with excutils.save_and_reraise_exception(): [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] self.force_reraise() [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] raise self.value [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] updated_port = self._update_port( [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] _ensure_no_port_binding_failure(port) [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.495378] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] raise exception.PortBindingFailed(port_id=port['id']) [ 658.495671] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] nova.exception.PortBindingFailed: Binding failed for port c84fc1ca-c5e3-40ba-a1d2-265501cf1467, please check neutron logs for more information. [ 658.495671] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] [ 658.495671] env[62521]: INFO nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Terminating instance [ 658.496166] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquiring lock "refresh_cache-fe69caab-dd71-4b01-a121-2591111b193d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.752743] env[62521]: DEBUG nova.network.neutron [req-b51c23d5-203b-45e3-89f6-2524d863ab48 req-2091e028-ad79-494e-b6b7-6f0692a1d675 service nova] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.757264] env[62521]: DEBUG nova.compute.utils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 658.764578] env[62521]: DEBUG nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 658.764828] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 658.835394] env[62521]: DEBUG nova.policy [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b961228193464ccbbe71d52cfca4d38c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f7c022ac207488c8478a6cb00aacb45', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 658.958955] env[62521]: DEBUG nova.network.neutron [req-b51c23d5-203b-45e3-89f6-2524d863ab48 req-2091e028-ad79-494e-b6b7-6f0692a1d675 service nova] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.263733] env[62521]: DEBUG nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 659.305340] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358bf7c5-1453-4f32-9c6e-5546a40d2575 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.315160] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a6455d-781e-4b11-8d6f-b48cc1e52311 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.354417] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d72c0f09-ce80-4df0-a106-127e82a43d6d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.363933] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1640a05-1cdf-4f7a-a9a9-fe6cf019bba1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.370138] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Successfully created port: 47f7f57d-d2dc-4ce6-8259-b7a863189fa7 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 659.382797] env[62521]: DEBUG nova.compute.provider_tree [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.464093] env[62521]: DEBUG oslo_concurrency.lockutils [req-b51c23d5-203b-45e3-89f6-2524d863ab48 req-2091e028-ad79-494e-b6b7-6f0692a1d675 service nova] Releasing lock "refresh_cache-fe69caab-dd71-4b01-a121-2591111b193d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.464536] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquired lock "refresh_cache-fe69caab-dd71-4b01-a121-2591111b193d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.464729] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.886811] env[62521]: DEBUG nova.scheduler.client.report [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.135994] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.279800] env[62521]: DEBUG nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 660.323018] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 660.323018] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 660.323018] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 660.323018] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 660.323422] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 660.323422] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 660.323422] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 660.323422] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 660.323422] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 660.323640] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 660.323640] env[62521]: DEBUG nova.virt.hardware [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 660.325131] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68250131-ffa7-4fec-9566-341990297fd5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.339025] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14934af5-23c2-4adf-8f52-aaf0f81c3dec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.356414] env[62521]: ERROR nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 47f7f57d-d2dc-4ce6-8259-b7a863189fa7, please check neutron logs for more information. [ 660.356414] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.356414] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.356414] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.356414] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.356414] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.356414] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.356414] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.356414] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.356414] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 660.356414] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.356414] env[62521]: ERROR nova.compute.manager raise self.value [ 660.356414] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.356414] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.356414] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.356414] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.356908] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.356908] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.356908] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 47f7f57d-d2dc-4ce6-8259-b7a863189fa7, please check neutron logs for more information. [ 660.356908] env[62521]: ERROR nova.compute.manager [ 660.356908] env[62521]: Traceback (most recent call last): [ 660.356908] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.356908] env[62521]: listener.cb(fileno) [ 660.356908] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.356908] env[62521]: result = function(*args, **kwargs) [ 660.356908] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.356908] env[62521]: return func(*args, **kwargs) [ 660.356908] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.356908] env[62521]: raise e [ 660.356908] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.356908] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 660.356908] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.356908] env[62521]: created_port_ids = self._update_ports_for_instance( [ 660.356908] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.356908] env[62521]: with excutils.save_and_reraise_exception(): [ 660.356908] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.356908] env[62521]: self.force_reraise() [ 660.356908] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.356908] env[62521]: raise self.value [ 660.356908] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.356908] env[62521]: updated_port = self._update_port( [ 660.356908] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.356908] env[62521]: _ensure_no_port_binding_failure(port) [ 660.356908] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.356908] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.357667] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 47f7f57d-d2dc-4ce6-8259-b7a863189fa7, please check neutron logs for more information. [ 660.357667] env[62521]: Removing descriptor: 18 [ 660.357667] env[62521]: ERROR nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 47f7f57d-d2dc-4ce6-8259-b7a863189fa7, please check neutron logs for more information. [ 660.357667] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Traceback (most recent call last): [ 660.357667] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 660.357667] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] yield resources [ 660.357667] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.357667] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] self.driver.spawn(context, instance, image_meta, [ 660.357667] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 660.357667] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.357667] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.357667] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] vm_ref = self.build_virtual_machine(instance, [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] for vif in network_info: [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] return self._sync_wrapper(fn, *args, **kwargs) [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] self.wait() [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] self[:] = self._gt.wait() [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] return self._exit_event.wait() [ 660.358009] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] result = hub.switch() [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] return self.greenlet.switch() [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] result = function(*args, **kwargs) [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] return func(*args, **kwargs) [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] raise e [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] nwinfo = self.network_api.allocate_for_instance( [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.358377] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] created_port_ids = self._update_ports_for_instance( [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] with excutils.save_and_reraise_exception(): [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] self.force_reraise() [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] raise self.value [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] updated_port = self._update_port( [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] _ensure_no_port_binding_failure(port) [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.358736] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] raise exception.PortBindingFailed(port_id=port['id']) [ 660.359081] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] nova.exception.PortBindingFailed: Binding failed for port 47f7f57d-d2dc-4ce6-8259-b7a863189fa7, please check neutron logs for more information. [ 660.359081] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] [ 660.359081] env[62521]: INFO nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Terminating instance [ 660.359903] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Acquiring lock "refresh_cache-edd89a57-f026-4a3e-8b6b-111cc2321d9d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.360074] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Acquired lock "refresh_cache-edd89a57-f026-4a3e-8b6b-111cc2321d9d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.360242] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.393024] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.141s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.393640] env[62521]: ERROR nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 876cdb7a-b71f-4b23-b194-e5cf93ec5691, please check neutron logs for more information. [ 660.393640] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Traceback (most recent call last): [ 660.393640] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.393640] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] self.driver.spawn(context, instance, image_meta, [ 660.393640] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 660.393640] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.393640] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.393640] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] vm_ref = self.build_virtual_machine(instance, [ 660.393640] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.393640] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.393640] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] for vif in network_info: [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] return self._sync_wrapper(fn, *args, **kwargs) [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] self.wait() [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] self[:] = self._gt.wait() [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] return self._exit_event.wait() [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] result = hub.switch() [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.393925] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] return self.greenlet.switch() [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] result = function(*args, **kwargs) [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] return func(*args, **kwargs) [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] raise e [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] nwinfo = self.network_api.allocate_for_instance( [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] created_port_ids = self._update_ports_for_instance( [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] with excutils.save_and_reraise_exception(): [ 660.394247] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] self.force_reraise() [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] raise self.value [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] updated_port = self._update_port( [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] _ensure_no_port_binding_failure(port) [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] raise exception.PortBindingFailed(port_id=port['id']) [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] nova.exception.PortBindingFailed: Binding failed for port 876cdb7a-b71f-4b23-b194-e5cf93ec5691, please check neutron logs for more information. [ 660.394577] env[62521]: ERROR nova.compute.manager [instance: d9571cc5-0555-4867-a8d3-a2281d512305] [ 660.394841] env[62521]: DEBUG nova.compute.utils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Binding failed for port 876cdb7a-b71f-4b23-b194-e5cf93ec5691, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.396439] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.905s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.396974] env[62521]: INFO nova.compute.claims [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.399583] env[62521]: DEBUG nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Build of instance d9571cc5-0555-4867-a8d3-a2281d512305 was re-scheduled: Binding failed for port 876cdb7a-b71f-4b23-b194-e5cf93ec5691, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 660.400411] env[62521]: DEBUG nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 660.400690] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquiring lock "refresh_cache-d9571cc5-0555-4867-a8d3-a2281d512305" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.400847] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquired lock "refresh_cache-d9571cc5-0555-4867-a8d3-a2281d512305" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.401017] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.481608] env[62521]: DEBUG nova.compute.manager [req-788389f5-941e-4b63-a17b-c6dab0de04ce req-831d8357-5d5a-45f7-a72b-a3062062a989 service nova] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Received event network-changed-47f7f57d-d2dc-4ce6-8259-b7a863189fa7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.481608] env[62521]: DEBUG nova.compute.manager [req-788389f5-941e-4b63-a17b-c6dab0de04ce req-831d8357-5d5a-45f7-a72b-a3062062a989 service nova] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Refreshing instance network info cache due to event network-changed-47f7f57d-d2dc-4ce6-8259-b7a863189fa7. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 660.481685] env[62521]: DEBUG oslo_concurrency.lockutils [req-788389f5-941e-4b63-a17b-c6dab0de04ce req-831d8357-5d5a-45f7-a72b-a3062062a989 service nova] Acquiring lock "refresh_cache-edd89a57-f026-4a3e-8b6b-111cc2321d9d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.508085] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.605520] env[62521]: DEBUG nova.compute.manager [req-a6d8ee00-a979-4b1b-aca5-e05491200cb8 req-fee22440-56dd-4a5e-8085-42a9d4471dbe service nova] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Received event network-vif-deleted-c84fc1ca-c5e3-40ba-a1d2-265501cf1467 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.927888] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.929077] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.013017] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Releasing lock "refresh_cache-fe69caab-dd71-4b01-a121-2591111b193d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.013017] env[62521]: DEBUG nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 661.013017] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 661.013017] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b576dc4-5cbc-42ac-aeba-85bda71727d5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.014954] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.025531] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8e5cc5-d9b1-45da-9ad7-e4e409aab174 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.056663] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fe69caab-dd71-4b01-a121-2591111b193d could not be found. [ 661.056663] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 661.056663] env[62521]: INFO nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 661.056663] env[62521]: DEBUG oslo.service.loopingcall [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 661.056663] env[62521]: DEBUG nova.compute.manager [-] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.056663] env[62521]: DEBUG nova.network.neutron [-] [instance: fe69caab-dd71-4b01-a121-2591111b193d] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.078688] env[62521]: DEBUG nova.network.neutron [-] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.227359] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.518297] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Releasing lock "refresh_cache-d9571cc5-0555-4867-a8d3-a2281d512305" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.523020] env[62521]: DEBUG nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 661.523020] env[62521]: DEBUG nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.523020] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.547628] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.582466] env[62521]: DEBUG nova.network.neutron [-] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.730526] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Releasing lock "refresh_cache-edd89a57-f026-4a3e-8b6b-111cc2321d9d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.734021] env[62521]: DEBUG nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 661.734021] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 661.734021] env[62521]: DEBUG oslo_concurrency.lockutils [req-788389f5-941e-4b63-a17b-c6dab0de04ce req-831d8357-5d5a-45f7-a72b-a3062062a989 service nova] Acquired lock "refresh_cache-edd89a57-f026-4a3e-8b6b-111cc2321d9d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.734021] env[62521]: DEBUG nova.network.neutron [req-788389f5-941e-4b63-a17b-c6dab0de04ce req-831d8357-5d5a-45f7-a72b-a3062062a989 service nova] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Refreshing network info cache for port 47f7f57d-d2dc-4ce6-8259-b7a863189fa7 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 661.734021] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e562f15-e0e0-4a1a-be9f-c04d881081a8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.748018] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26054cd9-897d-4975-a4a8-ca81d8a5f823 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.782245] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance edd89a57-f026-4a3e-8b6b-111cc2321d9d could not be found. [ 661.782630] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 661.782832] env[62521]: INFO nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 661.783140] env[62521]: DEBUG oslo.service.loopingcall [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 661.786896] env[62521]: DEBUG nova.compute.manager [-] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.786998] env[62521]: DEBUG nova.network.neutron [-] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.856084] env[62521]: DEBUG nova.network.neutron [-] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.970440] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9034ebd3-a862-4ef8-9fb4-844283a9466b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.983024] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371acd35-f58b-4e5f-950e-86594525bb23 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.016363] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-598e969c-36b0-4cb8-9945-8024165db5b5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.023969] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd46f356-a0c7-4be5-8d49-5b66e4690d02 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.039259] env[62521]: DEBUG nova.compute.provider_tree [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.053357] env[62521]: DEBUG nova.network.neutron [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.085812] env[62521]: INFO nova.compute.manager [-] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Took 1.03 seconds to deallocate network for instance. [ 662.088933] env[62521]: DEBUG nova.compute.claims [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 662.088933] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.265457] env[62521]: DEBUG nova.network.neutron [req-788389f5-941e-4b63-a17b-c6dab0de04ce req-831d8357-5d5a-45f7-a72b-a3062062a989 service nova] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.357694] env[62521]: DEBUG nova.network.neutron [-] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.420601] env[62521]: DEBUG nova.network.neutron [req-788389f5-941e-4b63-a17b-c6dab0de04ce req-831d8357-5d5a-45f7-a72b-a3062062a989 service nova] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.542217] env[62521]: DEBUG nova.scheduler.client.report [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.556430] env[62521]: INFO nova.compute.manager [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: d9571cc5-0555-4867-a8d3-a2281d512305] Took 1.04 seconds to deallocate network for instance. [ 662.802252] env[62521]: DEBUG nova.compute.manager [req-e8145870-0248-4675-9265-14d535497a61 req-ac845648-8b31-4321-a182-ad6c4df5209b service nova] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Received event network-vif-deleted-47f7f57d-d2dc-4ce6-8259-b7a863189fa7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.860799] env[62521]: INFO nova.compute.manager [-] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Took 1.07 seconds to deallocate network for instance. [ 662.864365] env[62521]: DEBUG nova.compute.claims [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 662.864365] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.924795] env[62521]: DEBUG oslo_concurrency.lockutils [req-788389f5-941e-4b63-a17b-c6dab0de04ce req-831d8357-5d5a-45f7-a72b-a3062062a989 service nova] Releasing lock "refresh_cache-edd89a57-f026-4a3e-8b6b-111cc2321d9d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.047997] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.652s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.047997] env[62521]: DEBUG nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 663.051943] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.339s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.057373] env[62521]: INFO nova.compute.claims [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.319736] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Acquiring lock "3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.323260] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Lock "3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.563997] env[62521]: DEBUG nova.compute.utils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.571928] env[62521]: DEBUG nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.572134] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 663.612321] env[62521]: INFO nova.scheduler.client.report [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Deleted allocations for instance d9571cc5-0555-4867-a8d3-a2281d512305 [ 663.670203] env[62521]: DEBUG nova.policy [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2acbdbd60b934cee9fcb1f8c79e3d644', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '01c5a629543a43cea19d0d14b5267919', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 664.077514] env[62521]: DEBUG nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 664.122608] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b157574c-180d-45a2-9bcc-2b8294f8225f tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "d9571cc5-0555-4867-a8d3-a2281d512305" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.062s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.197879] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Successfully created port: ef49cd51-3ad9-4500-a7fb-42b6cb8e8916 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.314186] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquiring lock "50aef23e-9429-4499-9fc1-c31b97e23029" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.314186] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lock "50aef23e-9429-4499-9fc1-c31b97e23029" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.494802] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.495267] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.632021] env[62521]: DEBUG nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 664.706435] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8016da6-6e50-40ce-aab2-b3c83f9c93ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.717428] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bee08d-c323-4d07-b55f-4c686b0debde {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.751929] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6b93bd-fd15-4346-aee9-0bdcb4e2f90c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.759914] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790cd8da-a504-462f-9139-f3685287ac66 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.774915] env[62521]: DEBUG nova.compute.provider_tree [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.096145] env[62521]: DEBUG nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 665.129292] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 665.129602] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 665.129766] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.129950] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 665.131066] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.131512] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 665.131817] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 665.132118] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 665.132361] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 665.132601] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 665.132840] env[62521]: DEBUG nova.virt.hardware [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 665.133852] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca678044-29bc-4960-aee3-5e3f38c99d1d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.147129] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e4f278-c0fe-4855-a2d0-f4634b7c4f12 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.161359] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.278785] env[62521]: DEBUG nova.scheduler.client.report [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.325043] env[62521]: DEBUG nova.compute.manager [req-5887cbe8-7ebd-41c6-a80f-e2c136b37407 req-516b3ebf-d8b0-4eeb-a30c-6db38b5243ba service nova] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Received event network-changed-ef49cd51-3ad9-4500-a7fb-42b6cb8e8916 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.325273] env[62521]: DEBUG nova.compute.manager [req-5887cbe8-7ebd-41c6-a80f-e2c136b37407 req-516b3ebf-d8b0-4eeb-a30c-6db38b5243ba service nova] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Refreshing instance network info cache due to event network-changed-ef49cd51-3ad9-4500-a7fb-42b6cb8e8916. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 665.325351] env[62521]: DEBUG oslo_concurrency.lockutils [req-5887cbe8-7ebd-41c6-a80f-e2c136b37407 req-516b3ebf-d8b0-4eeb-a30c-6db38b5243ba service nova] Acquiring lock "refresh_cache-104b64ff-a042-4bee-a5fd-bbbdf288f63f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.325481] env[62521]: DEBUG oslo_concurrency.lockutils [req-5887cbe8-7ebd-41c6-a80f-e2c136b37407 req-516b3ebf-d8b0-4eeb-a30c-6db38b5243ba service nova] Acquired lock "refresh_cache-104b64ff-a042-4bee-a5fd-bbbdf288f63f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.325670] env[62521]: DEBUG nova.network.neutron [req-5887cbe8-7ebd-41c6-a80f-e2c136b37407 req-516b3ebf-d8b0-4eeb-a30c-6db38b5243ba service nova] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Refreshing network info cache for port ef49cd51-3ad9-4500-a7fb-42b6cb8e8916 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 665.542298] env[62521]: ERROR nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ef49cd51-3ad9-4500-a7fb-42b6cb8e8916, please check neutron logs for more information. [ 665.542298] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 665.542298] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.542298] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 665.542298] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.542298] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 665.542298] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.542298] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 665.542298] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.542298] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 665.542298] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.542298] env[62521]: ERROR nova.compute.manager raise self.value [ 665.542298] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.542298] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 665.542298] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.542298] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 665.543076] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.543076] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 665.543076] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ef49cd51-3ad9-4500-a7fb-42b6cb8e8916, please check neutron logs for more information. [ 665.543076] env[62521]: ERROR nova.compute.manager [ 665.543076] env[62521]: Traceback (most recent call last): [ 665.543076] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 665.543076] env[62521]: listener.cb(fileno) [ 665.543076] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.543076] env[62521]: result = function(*args, **kwargs) [ 665.543076] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.543076] env[62521]: return func(*args, **kwargs) [ 665.543076] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.543076] env[62521]: raise e [ 665.543076] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.543076] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 665.543076] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.543076] env[62521]: created_port_ids = self._update_ports_for_instance( [ 665.543076] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.543076] env[62521]: with excutils.save_and_reraise_exception(): [ 665.543076] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.543076] env[62521]: self.force_reraise() [ 665.543076] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.543076] env[62521]: raise self.value [ 665.543076] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.543076] env[62521]: updated_port = self._update_port( [ 665.543076] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.543076] env[62521]: _ensure_no_port_binding_failure(port) [ 665.543076] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.543076] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 665.543799] env[62521]: nova.exception.PortBindingFailed: Binding failed for port ef49cd51-3ad9-4500-a7fb-42b6cb8e8916, please check neutron logs for more information. [ 665.543799] env[62521]: Removing descriptor: 15 [ 665.543799] env[62521]: ERROR nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ef49cd51-3ad9-4500-a7fb-42b6cb8e8916, please check neutron logs for more information. [ 665.543799] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Traceback (most recent call last): [ 665.543799] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.543799] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] yield resources [ 665.543799] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.543799] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] self.driver.spawn(context, instance, image_meta, [ 665.543799] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 665.543799] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.543799] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.543799] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] vm_ref = self.build_virtual_machine(instance, [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] for vif in network_info: [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] return self._sync_wrapper(fn, *args, **kwargs) [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] self.wait() [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] self[:] = self._gt.wait() [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] return self._exit_event.wait() [ 665.544134] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] result = hub.switch() [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] return self.greenlet.switch() [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] result = function(*args, **kwargs) [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] return func(*args, **kwargs) [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] raise e [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] nwinfo = self.network_api.allocate_for_instance( [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.544476] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] created_port_ids = self._update_ports_for_instance( [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] with excutils.save_and_reraise_exception(): [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] self.force_reraise() [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] raise self.value [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] updated_port = self._update_port( [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] _ensure_no_port_binding_failure(port) [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.544822] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] raise exception.PortBindingFailed(port_id=port['id']) [ 665.545178] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] nova.exception.PortBindingFailed: Binding failed for port ef49cd51-3ad9-4500-a7fb-42b6cb8e8916, please check neutron logs for more information. [ 665.545178] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] [ 665.545178] env[62521]: INFO nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Terminating instance [ 665.545942] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquiring lock "refresh_cache-104b64ff-a042-4bee-a5fd-bbbdf288f63f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.785096] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.733s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.785747] env[62521]: DEBUG nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 665.788485] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.123s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.789964] env[62521]: INFO nova.compute.claims [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 665.850936] env[62521]: DEBUG nova.network.neutron [req-5887cbe8-7ebd-41c6-a80f-e2c136b37407 req-516b3ebf-d8b0-4eeb-a30c-6db38b5243ba service nova] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.904700] env[62521]: DEBUG nova.network.neutron [req-5887cbe8-7ebd-41c6-a80f-e2c136b37407 req-516b3ebf-d8b0-4eeb-a30c-6db38b5243ba service nova] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.297018] env[62521]: DEBUG nova.compute.utils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.300966] env[62521]: DEBUG nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 666.300966] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 666.341136] env[62521]: DEBUG nova.policy [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac0171aba589411ab6a60481fccdfdff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f882ab7b2fb14655a5d09104b14e2bfc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 666.407463] env[62521]: DEBUG oslo_concurrency.lockutils [req-5887cbe8-7ebd-41c6-a80f-e2c136b37407 req-516b3ebf-d8b0-4eeb-a30c-6db38b5243ba service nova] Releasing lock "refresh_cache-104b64ff-a042-4bee-a5fd-bbbdf288f63f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.407917] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquired lock "refresh_cache-104b64ff-a042-4bee-a5fd-bbbdf288f63f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.408128] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 666.705520] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Successfully created port: 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 666.804428] env[62521]: DEBUG nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 666.927661] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.977452] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.307704] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a329717a-7b41-4e40-b590-7ebc97393539 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.324140] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b825ff-c1a8-4c31-b559-78d4b1a60db2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.362701] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd84db9-0ec8-42b4-a50d-b8d0d9d8374b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.371492] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a088a98-0e7f-441e-a4f2-0f2a9e0575ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.386250] env[62521]: DEBUG nova.compute.provider_tree [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.412316] env[62521]: DEBUG nova.compute.manager [req-0630a6fe-7fd7-4e09-8c1b-0091cf67ff4a req-a50d619b-1fd2-4052-a675-bc30c54ed1e4 service nova] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Received event network-vif-deleted-ef49cd51-3ad9-4500-a7fb-42b6cb8e8916 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 667.481954] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Releasing lock "refresh_cache-104b64ff-a042-4bee-a5fd-bbbdf288f63f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.482392] env[62521]: DEBUG nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.482611] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 667.482891] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9fb5a6dd-8257-427f-bd13-bb6b7a0db1e0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.491645] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4010d314-3f5c-49f3-bcc6-790bb2c27ef2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.512960] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 104b64ff-a042-4bee-a5fd-bbbdf288f63f could not be found. [ 667.513186] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 667.513374] env[62521]: INFO nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 667.513621] env[62521]: DEBUG oslo.service.loopingcall [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.513821] env[62521]: DEBUG nova.compute.manager [-] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.513914] env[62521]: DEBUG nova.network.neutron [-] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 667.528131] env[62521]: DEBUG nova.network.neutron [-] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.820679] env[62521]: DEBUG nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 667.856239] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 667.856524] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 667.856700] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 667.856886] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 667.857113] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 667.857207] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 667.857407] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 667.857655] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 667.857850] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 667.858026] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 667.858210] env[62521]: DEBUG nova.virt.hardware [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 667.859065] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490b3685-d618-4391-bb1b-f7ee529d86cf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.868700] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9dd03ca-8106-4c49-8c70-dcfbe7241df6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.888865] env[62521]: DEBUG nova.scheduler.client.report [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.034445] env[62521]: DEBUG nova.network.neutron [-] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.140837] env[62521]: ERROR nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6, please check neutron logs for more information. [ 668.140837] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 668.140837] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.140837] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 668.140837] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.140837] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 668.140837] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.140837] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 668.140837] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.140837] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 668.140837] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.140837] env[62521]: ERROR nova.compute.manager raise self.value [ 668.140837] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.140837] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 668.140837] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.140837] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 668.141520] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.141520] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 668.141520] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6, please check neutron logs for more information. [ 668.141520] env[62521]: ERROR nova.compute.manager [ 668.141520] env[62521]: Traceback (most recent call last): [ 668.141520] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 668.141520] env[62521]: listener.cb(fileno) [ 668.141520] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.141520] env[62521]: result = function(*args, **kwargs) [ 668.141520] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.141520] env[62521]: return func(*args, **kwargs) [ 668.141520] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.141520] env[62521]: raise e [ 668.141520] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.141520] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 668.141520] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.141520] env[62521]: created_port_ids = self._update_ports_for_instance( [ 668.141520] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.141520] env[62521]: with excutils.save_and_reraise_exception(): [ 668.141520] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.141520] env[62521]: self.force_reraise() [ 668.141520] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.141520] env[62521]: raise self.value [ 668.141520] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.141520] env[62521]: updated_port = self._update_port( [ 668.141520] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.141520] env[62521]: _ensure_no_port_binding_failure(port) [ 668.141520] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.141520] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 668.142394] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6, please check neutron logs for more information. [ 668.142394] env[62521]: Removing descriptor: 15 [ 668.142394] env[62521]: ERROR nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6, please check neutron logs for more information. [ 668.142394] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Traceback (most recent call last): [ 668.142394] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 668.142394] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] yield resources [ 668.142394] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.142394] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] self.driver.spawn(context, instance, image_meta, [ 668.142394] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 668.142394] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.142394] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.142394] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] vm_ref = self.build_virtual_machine(instance, [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] for vif in network_info: [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] return self._sync_wrapper(fn, *args, **kwargs) [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] self.wait() [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] self[:] = self._gt.wait() [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] return self._exit_event.wait() [ 668.142768] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] result = hub.switch() [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] return self.greenlet.switch() [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] result = function(*args, **kwargs) [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] return func(*args, **kwargs) [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] raise e [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] nwinfo = self.network_api.allocate_for_instance( [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.143167] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] created_port_ids = self._update_ports_for_instance( [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] with excutils.save_and_reraise_exception(): [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] self.force_reraise() [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] raise self.value [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] updated_port = self._update_port( [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] _ensure_no_port_binding_failure(port) [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.143536] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] raise exception.PortBindingFailed(port_id=port['id']) [ 668.143898] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] nova.exception.PortBindingFailed: Binding failed for port 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6, please check neutron logs for more information. [ 668.143898] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] [ 668.143898] env[62521]: INFO nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Terminating instance [ 668.144120] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Acquiring lock "refresh_cache-e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.144268] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Acquired lock "refresh_cache-e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.144436] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.393510] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.394029] env[62521]: DEBUG nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 668.396836] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 20.540s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.397014] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.397173] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 668.397442] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.748s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.398822] env[62521]: INFO nova.compute.claims [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 668.403958] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a41cf91-af0e-4796-a425-7d555f9a8d6a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.411681] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74fa80c-78f8-4848-8638-29de380b1d6d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.428079] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c00ff58-b388-46b8-9d39-4d4c17fe7ff5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.436274] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d9abe3-4bd3-42ad-8361-3681d28cfeb2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.466202] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181413MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 668.466366] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.537956] env[62521]: INFO nova.compute.manager [-] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Took 1.02 seconds to deallocate network for instance. [ 668.540386] env[62521]: DEBUG nova.compute.claims [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 668.540386] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.661807] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.769032] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.905212] env[62521]: DEBUG nova.compute.utils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 668.906889] env[62521]: DEBUG nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 668.907307] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 668.957807] env[62521]: DEBUG nova.policy [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0e99fc3ead94078b185bc599ff247ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25420ed4f82e478989a011d40f037a24', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 669.271460] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Releasing lock "refresh_cache-e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.271897] env[62521]: DEBUG nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 669.272101] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 669.272460] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51099d3f-7ccb-42a7-90d7-0a5e90bdb312 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.285081] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588659ea-9c33-4910-96bd-39ea27d7e2d9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.308720] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e0aabeb7-7d4c-4358-8f6c-e423a0baea3b could not be found. [ 669.308941] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 669.309491] env[62521]: INFO nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 669.309491] env[62521]: DEBUG oslo.service.loopingcall [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.309668] env[62521]: DEBUG nova.compute.manager [-] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.309730] env[62521]: DEBUG nova.network.neutron [-] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 669.327329] env[62521]: DEBUG nova.network.neutron [-] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.343287] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Successfully created port: 6cee9131-a87f-4722-88b3-00a7ba0882ae {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 669.412441] env[62521]: DEBUG nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 669.443984] env[62521]: DEBUG nova.compute.manager [req-a52c2f28-335f-4218-98b3-e35288723e3e req-dd3ca819-d31d-4931-8030-448275810d47 service nova] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Received event network-changed-3f54e9d5-40a8-4d3f-ab5f-0292a78229a6 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.444217] env[62521]: DEBUG nova.compute.manager [req-a52c2f28-335f-4218-98b3-e35288723e3e req-dd3ca819-d31d-4931-8030-448275810d47 service nova] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Refreshing instance network info cache due to event network-changed-3f54e9d5-40a8-4d3f-ab5f-0292a78229a6. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 669.444458] env[62521]: DEBUG oslo_concurrency.lockutils [req-a52c2f28-335f-4218-98b3-e35288723e3e req-dd3ca819-d31d-4931-8030-448275810d47 service nova] Acquiring lock "refresh_cache-e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.444713] env[62521]: DEBUG oslo_concurrency.lockutils [req-a52c2f28-335f-4218-98b3-e35288723e3e req-dd3ca819-d31d-4931-8030-448275810d47 service nova] Acquired lock "refresh_cache-e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.446261] env[62521]: DEBUG nova.network.neutron [req-a52c2f28-335f-4218-98b3-e35288723e3e req-dd3ca819-d31d-4931-8030-448275810d47 service nova] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Refreshing network info cache for port 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 669.833027] env[62521]: DEBUG nova.network.neutron [-] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.851106] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35778e07-6005-47a0-b1da-1d152d07e203 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.859438] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57d1bde-65be-4260-8047-d52723507328 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.889024] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96618f6b-fee2-4484-a541-3128459f16f5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.896375] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87cc4eb-9158-4e4a-80ae-01fea3649551 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.910761] env[62521]: DEBUG nova.compute.provider_tree [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.979406] env[62521]: DEBUG nova.network.neutron [req-a52c2f28-335f-4218-98b3-e35288723e3e req-dd3ca819-d31d-4931-8030-448275810d47 service nova] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.059661] env[62521]: DEBUG nova.network.neutron [req-a52c2f28-335f-4218-98b3-e35288723e3e req-dd3ca819-d31d-4931-8030-448275810d47 service nova] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.335591] env[62521]: INFO nova.compute.manager [-] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Took 1.03 seconds to deallocate network for instance. [ 670.337959] env[62521]: DEBUG nova.compute.claims [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 670.338156] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.414356] env[62521]: DEBUG nova.scheduler.client.report [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.424535] env[62521]: DEBUG nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 670.453887] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 670.454144] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 670.454305] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 670.454485] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 670.454663] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 670.454816] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 670.455084] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 670.455269] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 670.455443] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 670.455737] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 670.455864] env[62521]: DEBUG nova.virt.hardware [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 670.457481] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f0d9a0-9c88-4b8a-a028-3f5863a284b3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.466583] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c386058f-9c85-4c8f-a17c-e8397f3e72ae {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.518657] env[62521]: ERROR nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6cee9131-a87f-4722-88b3-00a7ba0882ae, please check neutron logs for more information. [ 670.518657] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 670.518657] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.518657] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 670.518657] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.518657] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 670.518657] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.518657] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 670.518657] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.518657] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 670.518657] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.518657] env[62521]: ERROR nova.compute.manager raise self.value [ 670.518657] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.518657] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 670.518657] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.518657] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 670.519526] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.519526] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 670.519526] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6cee9131-a87f-4722-88b3-00a7ba0882ae, please check neutron logs for more information. [ 670.519526] env[62521]: ERROR nova.compute.manager [ 670.519526] env[62521]: Traceback (most recent call last): [ 670.519526] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 670.519526] env[62521]: listener.cb(fileno) [ 670.519526] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.519526] env[62521]: result = function(*args, **kwargs) [ 670.519526] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 670.519526] env[62521]: return func(*args, **kwargs) [ 670.519526] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.519526] env[62521]: raise e [ 670.519526] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.519526] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 670.519526] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.519526] env[62521]: created_port_ids = self._update_ports_for_instance( [ 670.519526] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.519526] env[62521]: with excutils.save_and_reraise_exception(): [ 670.519526] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.519526] env[62521]: self.force_reraise() [ 670.519526] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.519526] env[62521]: raise self.value [ 670.519526] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.519526] env[62521]: updated_port = self._update_port( [ 670.519526] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.519526] env[62521]: _ensure_no_port_binding_failure(port) [ 670.519526] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.519526] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 670.520731] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 6cee9131-a87f-4722-88b3-00a7ba0882ae, please check neutron logs for more information. [ 670.520731] env[62521]: Removing descriptor: 15 [ 670.520731] env[62521]: ERROR nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6cee9131-a87f-4722-88b3-00a7ba0882ae, please check neutron logs for more information. [ 670.520731] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Traceback (most recent call last): [ 670.520731] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 670.520731] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] yield resources [ 670.520731] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.520731] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] self.driver.spawn(context, instance, image_meta, [ 670.520731] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 670.520731] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.520731] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.520731] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] vm_ref = self.build_virtual_machine(instance, [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] for vif in network_info: [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] return self._sync_wrapper(fn, *args, **kwargs) [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] self.wait() [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] self[:] = self._gt.wait() [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] return self._exit_event.wait() [ 670.521237] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] result = hub.switch() [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] return self.greenlet.switch() [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] result = function(*args, **kwargs) [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] return func(*args, **kwargs) [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] raise e [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] nwinfo = self.network_api.allocate_for_instance( [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.521748] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] created_port_ids = self._update_ports_for_instance( [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] with excutils.save_and_reraise_exception(): [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] self.force_reraise() [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] raise self.value [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] updated_port = self._update_port( [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] _ensure_no_port_binding_failure(port) [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.522317] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] raise exception.PortBindingFailed(port_id=port['id']) [ 670.522811] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] nova.exception.PortBindingFailed: Binding failed for port 6cee9131-a87f-4722-88b3-00a7ba0882ae, please check neutron logs for more information. [ 670.522811] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] [ 670.522811] env[62521]: INFO nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Terminating instance [ 670.522811] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "refresh_cache-215472ea-402c-4eb5-8ad3-b7ef756dfdc5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.522987] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "refresh_cache-215472ea-402c-4eb5-8ad3-b7ef756dfdc5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.523044] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.562437] env[62521]: DEBUG oslo_concurrency.lockutils [req-a52c2f28-335f-4218-98b3-e35288723e3e req-dd3ca819-d31d-4931-8030-448275810d47 service nova] Releasing lock "refresh_cache-e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.562437] env[62521]: DEBUG nova.compute.manager [req-a52c2f28-335f-4218-98b3-e35288723e3e req-dd3ca819-d31d-4931-8030-448275810d47 service nova] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Received event network-vif-deleted-3f54e9d5-40a8-4d3f-ab5f-0292a78229a6 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.923019] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.923019] env[62521]: DEBUG nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 670.923019] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.967s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.924643] env[62521]: INFO nova.compute.claims [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.039954] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.130922] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.430278] env[62521]: DEBUG nova.compute.utils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 671.435448] env[62521]: DEBUG nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 671.435448] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 671.487955] env[62521]: DEBUG nova.compute.manager [req-6c38b8a8-f84f-4973-a992-601ac13667a8 req-5b67e231-8a40-455f-8146-c3e23d4d07bd service nova] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Received event network-changed-6cee9131-a87f-4722-88b3-00a7ba0882ae {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.488174] env[62521]: DEBUG nova.compute.manager [req-6c38b8a8-f84f-4973-a992-601ac13667a8 req-5b67e231-8a40-455f-8146-c3e23d4d07bd service nova] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Refreshing instance network info cache due to event network-changed-6cee9131-a87f-4722-88b3-00a7ba0882ae. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 671.488363] env[62521]: DEBUG oslo_concurrency.lockutils [req-6c38b8a8-f84f-4973-a992-601ac13667a8 req-5b67e231-8a40-455f-8146-c3e23d4d07bd service nova] Acquiring lock "refresh_cache-215472ea-402c-4eb5-8ad3-b7ef756dfdc5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.495538] env[62521]: DEBUG nova.policy [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d7aab6f3f6245c2b56979677428a9df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e6baab1556e4b00b07ab554664b4445', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 671.633220] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "refresh_cache-215472ea-402c-4eb5-8ad3-b7ef756dfdc5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.633648] env[62521]: DEBUG nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.633901] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 671.634248] env[62521]: DEBUG oslo_concurrency.lockutils [req-6c38b8a8-f84f-4973-a992-601ac13667a8 req-5b67e231-8a40-455f-8146-c3e23d4d07bd service nova] Acquired lock "refresh_cache-215472ea-402c-4eb5-8ad3-b7ef756dfdc5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.634420] env[62521]: DEBUG nova.network.neutron [req-6c38b8a8-f84f-4973-a992-601ac13667a8 req-5b67e231-8a40-455f-8146-c3e23d4d07bd service nova] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Refreshing network info cache for port 6cee9131-a87f-4722-88b3-00a7ba0882ae {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 671.635575] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2130a5f8-9585-490b-b5d6-58367e194106 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.645434] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c699cfb6-9435-4b84-9ea2-c4b3bb95697f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.669313] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 215472ea-402c-4eb5-8ad3-b7ef756dfdc5 could not be found. [ 671.669422] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 671.669650] env[62521]: INFO nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 671.669902] env[62521]: DEBUG oslo.service.loopingcall [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.670131] env[62521]: DEBUG nova.compute.manager [-] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.670226] env[62521]: DEBUG nova.network.neutron [-] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.685896] env[62521]: DEBUG nova.network.neutron [-] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.814529] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Successfully created port: c9f4099c-087b-4d78-b799-e54699bcfcd9 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 671.935919] env[62521]: DEBUG nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 672.162838] env[62521]: DEBUG nova.network.neutron [req-6c38b8a8-f84f-4973-a992-601ac13667a8 req-5b67e231-8a40-455f-8146-c3e23d4d07bd service nova] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.188940] env[62521]: DEBUG nova.network.neutron [-] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.391562] env[62521]: DEBUG nova.network.neutron [req-6c38b8a8-f84f-4973-a992-601ac13667a8 req-5b67e231-8a40-455f-8146-c3e23d4d07bd service nova] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.421129] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7485120d-72bf-4f52-8ee2-88a3663a0d3f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.429356] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d53ab63-edfd-4f36-8a41-eee023f2e8fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.462179] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d63a88-0fb3-4070-9b91-8d52f886317f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.469732] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69db81c1-3894-4e8b-bef3-283309a2e039 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.484092] env[62521]: DEBUG nova.compute.provider_tree [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.691472] env[62521]: INFO nova.compute.manager [-] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Took 1.02 seconds to deallocate network for instance. [ 672.693897] env[62521]: DEBUG nova.compute.claims [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 672.694080] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.894723] env[62521]: DEBUG oslo_concurrency.lockutils [req-6c38b8a8-f84f-4973-a992-601ac13667a8 req-5b67e231-8a40-455f-8146-c3e23d4d07bd service nova] Releasing lock "refresh_cache-215472ea-402c-4eb5-8ad3-b7ef756dfdc5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.895040] env[62521]: DEBUG nova.compute.manager [req-6c38b8a8-f84f-4973-a992-601ac13667a8 req-5b67e231-8a40-455f-8146-c3e23d4d07bd service nova] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Received event network-vif-deleted-6cee9131-a87f-4722-88b3-00a7ba0882ae {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 672.966574] env[62521]: DEBUG nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 672.992830] env[62521]: DEBUG nova.scheduler.client.report [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.004498] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 673.004770] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 673.004930] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 673.005173] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 673.005332] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 673.005482] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 673.005971] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 673.005971] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 673.006118] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 673.006232] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 673.006411] env[62521]: DEBUG nova.virt.hardware [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 673.007298] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1f8d5d-069b-45b8-a509-0d85a2008538 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.016011] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d03ab8-0d4b-43e5-9a0b-1d00288ec1f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.165462] env[62521]: ERROR nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c9f4099c-087b-4d78-b799-e54699bcfcd9, please check neutron logs for more information. [ 673.165462] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 673.165462] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.165462] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 673.165462] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.165462] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 673.165462] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.165462] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 673.165462] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.165462] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 673.165462] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.165462] env[62521]: ERROR nova.compute.manager raise self.value [ 673.165462] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.165462] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 673.165462] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.165462] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 673.165947] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.165947] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 673.165947] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c9f4099c-087b-4d78-b799-e54699bcfcd9, please check neutron logs for more information. [ 673.165947] env[62521]: ERROR nova.compute.manager [ 673.165947] env[62521]: Traceback (most recent call last): [ 673.165947] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 673.165947] env[62521]: listener.cb(fileno) [ 673.165947] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.165947] env[62521]: result = function(*args, **kwargs) [ 673.165947] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 673.165947] env[62521]: return func(*args, **kwargs) [ 673.165947] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 673.165947] env[62521]: raise e [ 673.165947] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.165947] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 673.165947] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.165947] env[62521]: created_port_ids = self._update_ports_for_instance( [ 673.165947] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.165947] env[62521]: with excutils.save_and_reraise_exception(): [ 673.165947] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.165947] env[62521]: self.force_reraise() [ 673.165947] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.165947] env[62521]: raise self.value [ 673.165947] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.165947] env[62521]: updated_port = self._update_port( [ 673.165947] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.165947] env[62521]: _ensure_no_port_binding_failure(port) [ 673.165947] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.165947] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 673.166769] env[62521]: nova.exception.PortBindingFailed: Binding failed for port c9f4099c-087b-4d78-b799-e54699bcfcd9, please check neutron logs for more information. [ 673.166769] env[62521]: Removing descriptor: 15 [ 673.166769] env[62521]: ERROR nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c9f4099c-087b-4d78-b799-e54699bcfcd9, please check neutron logs for more information. [ 673.166769] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Traceback (most recent call last): [ 673.166769] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 673.166769] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] yield resources [ 673.166769] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 673.166769] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] self.driver.spawn(context, instance, image_meta, [ 673.166769] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 673.166769] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 673.166769] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 673.166769] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] vm_ref = self.build_virtual_machine(instance, [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] for vif in network_info: [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] return self._sync_wrapper(fn, *args, **kwargs) [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] self.wait() [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] self[:] = self._gt.wait() [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] return self._exit_event.wait() [ 673.167097] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] result = hub.switch() [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] return self.greenlet.switch() [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] result = function(*args, **kwargs) [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] return func(*args, **kwargs) [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] raise e [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] nwinfo = self.network_api.allocate_for_instance( [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.167597] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] created_port_ids = self._update_ports_for_instance( [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] with excutils.save_and_reraise_exception(): [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] self.force_reraise() [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] raise self.value [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] updated_port = self._update_port( [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] _ensure_no_port_binding_failure(port) [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.167930] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] raise exception.PortBindingFailed(port_id=port['id']) [ 673.168299] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] nova.exception.PortBindingFailed: Binding failed for port c9f4099c-087b-4d78-b799-e54699bcfcd9, please check neutron logs for more information. [ 673.168299] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] [ 673.168299] env[62521]: INFO nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Terminating instance [ 673.169523] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-f7611f03-73bb-4c0c-b787-d3c7f077f5bf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.169829] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-f7611f03-73bb-4c0c-b787-d3c7f077f5bf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.169909] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 673.503594] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.581s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.503792] env[62521]: DEBUG nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 673.507426] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.738s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.508660] env[62521]: INFO nova.compute.claims [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.590872] env[62521]: DEBUG nova.compute.manager [req-203f7d02-8d47-464a-99d8-6a8f7d0ad1f6 req-6a5f54f8-c209-4246-8c59-f5907d65943e service nova] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Received event network-changed-c9f4099c-087b-4d78-b799-e54699bcfcd9 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 673.591127] env[62521]: DEBUG nova.compute.manager [req-203f7d02-8d47-464a-99d8-6a8f7d0ad1f6 req-6a5f54f8-c209-4246-8c59-f5907d65943e service nova] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Refreshing instance network info cache due to event network-changed-c9f4099c-087b-4d78-b799-e54699bcfcd9. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 673.591286] env[62521]: DEBUG oslo_concurrency.lockutils [req-203f7d02-8d47-464a-99d8-6a8f7d0ad1f6 req-6a5f54f8-c209-4246-8c59-f5907d65943e service nova] Acquiring lock "refresh_cache-f7611f03-73bb-4c0c-b787-d3c7f077f5bf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.695578] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.804322] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.013315] env[62521]: DEBUG nova.compute.utils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 674.016643] env[62521]: DEBUG nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 674.017238] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 674.082884] env[62521]: DEBUG nova.policy [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f014b326172f4c958dcac8cddc519193', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9312e294a9424f0099432a60240d7ac2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 674.307771] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-f7611f03-73bb-4c0c-b787-d3c7f077f5bf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.308383] env[62521]: DEBUG nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 674.309331] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 674.309331] env[62521]: DEBUG oslo_concurrency.lockutils [req-203f7d02-8d47-464a-99d8-6a8f7d0ad1f6 req-6a5f54f8-c209-4246-8c59-f5907d65943e service nova] Acquired lock "refresh_cache-f7611f03-73bb-4c0c-b787-d3c7f077f5bf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.309331] env[62521]: DEBUG nova.network.neutron [req-203f7d02-8d47-464a-99d8-6a8f7d0ad1f6 req-6a5f54f8-c209-4246-8c59-f5907d65943e service nova] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Refreshing network info cache for port c9f4099c-087b-4d78-b799-e54699bcfcd9 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 674.313022] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb299a11-1663-4748-9801-7dc88537afa3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.322333] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d825fd8a-fc2f-412c-a2e6-453d486e3ac5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.348497] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f7611f03-73bb-4c0c-b787-d3c7f077f5bf could not be found. [ 674.348769] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 674.349054] env[62521]: INFO nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 674.349319] env[62521]: DEBUG oslo.service.loopingcall [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 674.349575] env[62521]: DEBUG nova.compute.manager [-] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.349701] env[62521]: DEBUG nova.network.neutron [-] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 674.371380] env[62521]: DEBUG nova.network.neutron [-] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.519511] env[62521]: DEBUG nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.533195] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Acquiring lock "e8cabcdc-d728-434d-a40d-fe751db086c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.533195] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "e8cabcdc-d728-434d-a40d-fe751db086c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.575363] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Successfully created port: 3fce0f5c-936f-4ca0-94ca-d2cb13d63094 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 674.854254] env[62521]: DEBUG nova.network.neutron [req-203f7d02-8d47-464a-99d8-6a8f7d0ad1f6 req-6a5f54f8-c209-4246-8c59-f5907d65943e service nova] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.874947] env[62521]: DEBUG nova.network.neutron [-] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.951895] env[62521]: DEBUG nova.network.neutron [req-203f7d02-8d47-464a-99d8-6a8f7d0ad1f6 req-6a5f54f8-c209-4246-8c59-f5907d65943e service nova] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.032709] env[62521]: INFO nova.virt.block_device [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Booting with volume 7853bc84-b6fa-4d53-b7b0-b82c12e612d2 at /dev/sda [ 675.051093] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f83e3f-90b4-4c4c-a4aa-e9761dd0e221 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.060901] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7792c03-a0b8-4ca7-bef1-0a4915c250f7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.095738] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40d64f3-30e9-4042-9a31-bbb64e6b7628 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.098388] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a16ba38f-fd24-4dc0-ae99-c7d8e9865d42 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.108277] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46315bd2-85ba-45ea-abc5-72f21873846f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.114637] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1abd0e6c-dee4-441e-a01a-012478160e91 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.136816] env[62521]: DEBUG nova.compute.provider_tree [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.142327] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b800bfea-48e4-472d-964f-b588908cfd53 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.152635] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6075dec-fcde-4a07-bc4b-3ea4f2dcb0fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.180044] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425da0f8-2e17-4d01-88c6-6e8f659de8b6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.188890] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586a2819-1b57-4355-b084-ad1de00cc6d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.206510] env[62521]: DEBUG nova.virt.block_device [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Updating existing volume attachment record: 5ca9e46d-37b6-4498-a4a1-1c1c83cb53ff {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 675.379547] env[62521]: INFO nova.compute.manager [-] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Took 1.03 seconds to deallocate network for instance. [ 675.382476] env[62521]: DEBUG nova.compute.claims [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 675.382763] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.455807] env[62521]: DEBUG oslo_concurrency.lockutils [req-203f7d02-8d47-464a-99d8-6a8f7d0ad1f6 req-6a5f54f8-c209-4246-8c59-f5907d65943e service nova] Releasing lock "refresh_cache-f7611f03-73bb-4c0c-b787-d3c7f077f5bf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.455807] env[62521]: DEBUG nova.compute.manager [req-203f7d02-8d47-464a-99d8-6a8f7d0ad1f6 req-6a5f54f8-c209-4246-8c59-f5907d65943e service nova] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Received event network-vif-deleted-c9f4099c-087b-4d78-b799-e54699bcfcd9 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.643923] env[62521]: DEBUG nova.scheduler.client.report [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.719843] env[62521]: DEBUG nova.compute.manager [req-ab22e32d-df06-4037-a512-082a4247dacc req-8189abcf-4195-49e5-8846-b3891bbc0ee7 service nova] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Received event network-changed-3fce0f5c-936f-4ca0-94ca-d2cb13d63094 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.720106] env[62521]: DEBUG nova.compute.manager [req-ab22e32d-df06-4037-a512-082a4247dacc req-8189abcf-4195-49e5-8846-b3891bbc0ee7 service nova] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Refreshing instance network info cache due to event network-changed-3fce0f5c-936f-4ca0-94ca-d2cb13d63094. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 675.720337] env[62521]: DEBUG oslo_concurrency.lockutils [req-ab22e32d-df06-4037-a512-082a4247dacc req-8189abcf-4195-49e5-8846-b3891bbc0ee7 service nova] Acquiring lock "refresh_cache-e1ead278-d904-41cc-bdfe-65174e8d6816" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.720480] env[62521]: DEBUG oslo_concurrency.lockutils [req-ab22e32d-df06-4037-a512-082a4247dacc req-8189abcf-4195-49e5-8846-b3891bbc0ee7 service nova] Acquired lock "refresh_cache-e1ead278-d904-41cc-bdfe-65174e8d6816" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.720639] env[62521]: DEBUG nova.network.neutron [req-ab22e32d-df06-4037-a512-082a4247dacc req-8189abcf-4195-49e5-8846-b3891bbc0ee7 service nova] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Refreshing network info cache for port 3fce0f5c-936f-4ca0-94ca-d2cb13d63094 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 675.950795] env[62521]: ERROR nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3fce0f5c-936f-4ca0-94ca-d2cb13d63094, please check neutron logs for more information. [ 675.950795] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 675.950795] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.950795] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 675.950795] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.950795] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 675.950795] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.950795] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 675.950795] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.950795] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 675.950795] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.950795] env[62521]: ERROR nova.compute.manager raise self.value [ 675.950795] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.950795] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 675.950795] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.950795] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 675.951275] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.951275] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 675.951275] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3fce0f5c-936f-4ca0-94ca-d2cb13d63094, please check neutron logs for more information. [ 675.951275] env[62521]: ERROR nova.compute.manager [ 675.951275] env[62521]: Traceback (most recent call last): [ 675.951275] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 675.951275] env[62521]: listener.cb(fileno) [ 675.951275] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.951275] env[62521]: result = function(*args, **kwargs) [ 675.951275] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 675.951275] env[62521]: return func(*args, **kwargs) [ 675.951275] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.951275] env[62521]: raise e [ 675.951275] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.951275] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 675.951275] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.951275] env[62521]: created_port_ids = self._update_ports_for_instance( [ 675.951275] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.951275] env[62521]: with excutils.save_and_reraise_exception(): [ 675.951275] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.951275] env[62521]: self.force_reraise() [ 675.951275] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.951275] env[62521]: raise self.value [ 675.951275] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.951275] env[62521]: updated_port = self._update_port( [ 675.951275] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.951275] env[62521]: _ensure_no_port_binding_failure(port) [ 675.951275] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.951275] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 675.953307] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 3fce0f5c-936f-4ca0-94ca-d2cb13d63094, please check neutron logs for more information. [ 675.953307] env[62521]: Removing descriptor: 15 [ 676.150029] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.642s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.150029] env[62521]: DEBUG nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 676.152751] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.348s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.154330] env[62521]: INFO nova.compute.claims [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.244099] env[62521]: DEBUG nova.network.neutron [req-ab22e32d-df06-4037-a512-082a4247dacc req-8189abcf-4195-49e5-8846-b3891bbc0ee7 service nova] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.347398] env[62521]: DEBUG nova.network.neutron [req-ab22e32d-df06-4037-a512-082a4247dacc req-8189abcf-4195-49e5-8846-b3891bbc0ee7 service nova] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.663658] env[62521]: DEBUG nova.compute.utils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 676.665736] env[62521]: DEBUG nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 676.666165] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 676.721647] env[62521]: DEBUG nova.policy [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73b4a00fafcb4eb588766c67dca79ab3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '231c01a1e06e401dabdd9ccb42a1697b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 676.849534] env[62521]: DEBUG oslo_concurrency.lockutils [req-ab22e32d-df06-4037-a512-082a4247dacc req-8189abcf-4195-49e5-8846-b3891bbc0ee7 service nova] Releasing lock "refresh_cache-e1ead278-d904-41cc-bdfe-65174e8d6816" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.106366] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Successfully created port: fc7e5ec2-ec04-441b-86c3-97b6288e0367 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 677.166688] env[62521]: DEBUG nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 677.330373] env[62521]: DEBUG nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.330942] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.331205] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.331394] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.331616] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.331811] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.331994] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.332287] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.332432] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.332671] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.332890] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.333122] env[62521]: DEBUG nova.virt.hardware [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.334337] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce4e8c21-32be-4aed-b1ae-142899585ac9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.343225] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0619f2a7-cd61-449a-8df3-fbc2f0b9518a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.364300] env[62521]: ERROR nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3fce0f5c-936f-4ca0-94ca-d2cb13d63094, please check neutron logs for more information. [ 677.364300] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Traceback (most recent call last): [ 677.364300] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 677.364300] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] yield resources [ 677.364300] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.364300] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] self.driver.spawn(context, instance, image_meta, [ 677.364300] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 677.364300] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.364300] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.364300] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] vm_ref = self.build_virtual_machine(instance, [ 677.364300] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] for vif in network_info: [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] return self._sync_wrapper(fn, *args, **kwargs) [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] self.wait() [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] self[:] = self._gt.wait() [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] return self._exit_event.wait() [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.364686] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] current.throw(*self._exc) [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] result = function(*args, **kwargs) [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] return func(*args, **kwargs) [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] raise e [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] nwinfo = self.network_api.allocate_for_instance( [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] created_port_ids = self._update_ports_for_instance( [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] with excutils.save_and_reraise_exception(): [ 677.365013] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] self.force_reraise() [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] raise self.value [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] updated_port = self._update_port( [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] _ensure_no_port_binding_failure(port) [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] raise exception.PortBindingFailed(port_id=port['id']) [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] nova.exception.PortBindingFailed: Binding failed for port 3fce0f5c-936f-4ca0-94ca-d2cb13d63094, please check neutron logs for more information. [ 677.365327] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] [ 677.365327] env[62521]: INFO nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Terminating instance [ 677.369132] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Acquiring lock "refresh_cache-e1ead278-d904-41cc-bdfe-65174e8d6816" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.369132] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Acquired lock "refresh_cache-e1ead278-d904-41cc-bdfe-65174e8d6816" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.369132] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.594820] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b88137-0f5d-458a-a530-2d0a54834125 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.604329] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1d5c47-d000-4184-9c96-cf05dc034c10 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.637118] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b3c009-63a4-4228-8f8a-88f9cc03477a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.644210] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ef1969-ceb5-4955-9caa-d4ad9b919f9e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.659761] env[62521]: DEBUG nova.compute.provider_tree [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.784540] env[62521]: DEBUG nova.compute.manager [req-a3b61e41-6f1d-4e0a-a99b-755d5bd6e3c6 req-418fe3ec-9720-4b78-ac0a-22deaec6474b service nova] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Received event network-vif-deleted-3fce0f5c-936f-4ca0-94ca-d2cb13d63094 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.892859] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.970233] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.165929] env[62521]: DEBUG nova.scheduler.client.report [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.179363] env[62521]: DEBUG nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 678.209801] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 678.210044] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 678.210210] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 678.210395] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 678.210543] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 678.210689] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 678.210939] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 678.211145] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 678.211353] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 678.211522] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 678.211693] env[62521]: DEBUG nova.virt.hardware [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 678.212643] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316c25de-0a13-435f-aaa4-5d0b567267f8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.221044] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbe795c-3674-43a1-87a3-d84713fa770d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.271937] env[62521]: ERROR nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fc7e5ec2-ec04-441b-86c3-97b6288e0367, please check neutron logs for more information. [ 678.271937] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 678.271937] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.271937] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 678.271937] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.271937] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 678.271937] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.271937] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 678.271937] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.271937] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 678.271937] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.271937] env[62521]: ERROR nova.compute.manager raise self.value [ 678.271937] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.271937] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 678.271937] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.271937] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 678.272463] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.272463] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 678.272463] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fc7e5ec2-ec04-441b-86c3-97b6288e0367, please check neutron logs for more information. [ 678.272463] env[62521]: ERROR nova.compute.manager [ 678.272463] env[62521]: Traceback (most recent call last): [ 678.272463] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 678.272463] env[62521]: listener.cb(fileno) [ 678.272463] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.272463] env[62521]: result = function(*args, **kwargs) [ 678.272463] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 678.272463] env[62521]: return func(*args, **kwargs) [ 678.272463] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.272463] env[62521]: raise e [ 678.272463] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.272463] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 678.272463] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.272463] env[62521]: created_port_ids = self._update_ports_for_instance( [ 678.272463] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.272463] env[62521]: with excutils.save_and_reraise_exception(): [ 678.272463] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.272463] env[62521]: self.force_reraise() [ 678.272463] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.272463] env[62521]: raise self.value [ 678.272463] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.272463] env[62521]: updated_port = self._update_port( [ 678.272463] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.272463] env[62521]: _ensure_no_port_binding_failure(port) [ 678.272463] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.272463] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 678.273302] env[62521]: nova.exception.PortBindingFailed: Binding failed for port fc7e5ec2-ec04-441b-86c3-97b6288e0367, please check neutron logs for more information. [ 678.273302] env[62521]: Removing descriptor: 15 [ 678.273302] env[62521]: ERROR nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fc7e5ec2-ec04-441b-86c3-97b6288e0367, please check neutron logs for more information. [ 678.273302] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Traceback (most recent call last): [ 678.273302] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 678.273302] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] yield resources [ 678.273302] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 678.273302] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] self.driver.spawn(context, instance, image_meta, [ 678.273302] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 678.273302] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 678.273302] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 678.273302] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] vm_ref = self.build_virtual_machine(instance, [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] vif_infos = vmwarevif.get_vif_info(self._session, [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] for vif in network_info: [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] return self._sync_wrapper(fn, *args, **kwargs) [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] self.wait() [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] self[:] = self._gt.wait() [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] return self._exit_event.wait() [ 678.273652] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] result = hub.switch() [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] return self.greenlet.switch() [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] result = function(*args, **kwargs) [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] return func(*args, **kwargs) [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] raise e [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] nwinfo = self.network_api.allocate_for_instance( [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.274017] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] created_port_ids = self._update_ports_for_instance( [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] with excutils.save_and_reraise_exception(): [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] self.force_reraise() [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] raise self.value [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] updated_port = self._update_port( [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] _ensure_no_port_binding_failure(port) [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.274363] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] raise exception.PortBindingFailed(port_id=port['id']) [ 678.274653] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] nova.exception.PortBindingFailed: Binding failed for port fc7e5ec2-ec04-441b-86c3-97b6288e0367, please check neutron logs for more information. [ 678.274653] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] [ 678.274653] env[62521]: INFO nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Terminating instance [ 678.278756] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Acquiring lock "refresh_cache-166786ee-4b89-4683-831b-9f3c1c57fb08" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.278920] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Acquired lock "refresh_cache-166786ee-4b89-4683-831b-9f3c1c57fb08" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.279101] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.472874] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Releasing lock "refresh_cache-e1ead278-d904-41cc-bdfe-65174e8d6816" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.473569] env[62521]: DEBUG nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 678.473823] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6bffbb0d-4896-4d3f-a365-2771255afd3b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.482930] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca1a454-8fa4-4715-b86c-a0a59f89bd4f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.505077] env[62521]: WARNING nova.virt.vmwareapi.driver [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance e1ead278-d904-41cc-bdfe-65174e8d6816 could not be found. [ 678.505394] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 678.505756] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04ed22f1-251d-463f-8c00-7f037b529c11 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.513389] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94275e79-40c1-43c5-ad9d-0a6b868a1b59 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.539148] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e1ead278-d904-41cc-bdfe-65174e8d6816 could not be found. [ 678.539148] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 678.539148] env[62521]: INFO nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Took 0.06 seconds to destroy the instance on the hypervisor. [ 678.539148] env[62521]: DEBUG oslo.service.loopingcall [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.539148] env[62521]: DEBUG nova.compute.manager [-] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.539148] env[62521]: DEBUG nova.network.neutron [-] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 678.552924] env[62521]: DEBUG nova.network.neutron [-] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.673781] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.674347] env[62521]: DEBUG nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 678.677630] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.588s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.797388] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.913789] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.055343] env[62521]: DEBUG nova.network.neutron [-] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.182203] env[62521]: DEBUG nova.compute.utils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.187208] env[62521]: DEBUG nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 679.187317] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 679.231707] env[62521]: DEBUG nova.policy [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a25805dff5749a99c308010a121a87a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a8a8a0b18e24eb0b8edb265bf018996', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.416495] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Releasing lock "refresh_cache-166786ee-4b89-4683-831b-9f3c1c57fb08" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.417188] env[62521]: DEBUG nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.417188] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 679.417479] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f2ea5150-455b-4053-a7d9-aa733c480f32 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.426353] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5f049d-fb33-44a7-8faa-3c55c361a1f5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.451430] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 166786ee-4b89-4683-831b-9f3c1c57fb08 could not be found. [ 679.451657] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 679.451840] env[62521]: INFO nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Took 0.03 seconds to destroy the instance on the hypervisor. [ 679.452091] env[62521]: DEBUG oslo.service.loopingcall [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.454448] env[62521]: DEBUG nova.compute.manager [-] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.454543] env[62521]: DEBUG nova.network.neutron [-] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.473258] env[62521]: DEBUG nova.network.neutron [-] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.560469] env[62521]: INFO nova.compute.manager [-] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Took 1.02 seconds to deallocate network for instance. [ 679.616353] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Successfully created port: f66095d6-0bf8-4cfc-9921-8fda94e576ea {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.620419] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68bb78b6-75e0-4ac1-99ae-b8612001566c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.628932] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221c49de-3181-452c-8e3e-cd6f0e4b3927 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.660965] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e786e4b-1981-497b-9d43-80c12432f0a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.668914] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f411c41-2e2c-45cc-9ff9-9466bfc775bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.682125] env[62521]: DEBUG nova.compute.provider_tree [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.687482] env[62521]: DEBUG nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 679.842471] env[62521]: DEBUG nova.compute.manager [req-0735d331-1735-44b9-bb6f-eb6e819354d4 req-dfac5646-268f-4bf0-882c-643ee2071b51 service nova] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Received event network-changed-fc7e5ec2-ec04-441b-86c3-97b6288e0367 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 679.843679] env[62521]: DEBUG nova.compute.manager [req-0735d331-1735-44b9-bb6f-eb6e819354d4 req-dfac5646-268f-4bf0-882c-643ee2071b51 service nova] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Refreshing instance network info cache due to event network-changed-fc7e5ec2-ec04-441b-86c3-97b6288e0367. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 679.843679] env[62521]: DEBUG oslo_concurrency.lockutils [req-0735d331-1735-44b9-bb6f-eb6e819354d4 req-dfac5646-268f-4bf0-882c-643ee2071b51 service nova] Acquiring lock "refresh_cache-166786ee-4b89-4683-831b-9f3c1c57fb08" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.843679] env[62521]: DEBUG oslo_concurrency.lockutils [req-0735d331-1735-44b9-bb6f-eb6e819354d4 req-dfac5646-268f-4bf0-882c-643ee2071b51 service nova] Acquired lock "refresh_cache-166786ee-4b89-4683-831b-9f3c1c57fb08" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.843679] env[62521]: DEBUG nova.network.neutron [req-0735d331-1735-44b9-bb6f-eb6e819354d4 req-dfac5646-268f-4bf0-882c-643ee2071b51 service nova] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Refreshing network info cache for port fc7e5ec2-ec04-441b-86c3-97b6288e0367 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 679.977345] env[62521]: DEBUG nova.network.neutron [-] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.125341] env[62521]: INFO nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Took 0.56 seconds to detach 1 volumes for instance. [ 680.129585] env[62521]: DEBUG nova.compute.claims [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 680.129807] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.187888] env[62521]: DEBUG nova.scheduler.client.report [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.362745] env[62521]: DEBUG nova.network.neutron [req-0735d331-1735-44b9-bb6f-eb6e819354d4 req-dfac5646-268f-4bf0-882c-643ee2071b51 service nova] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.440859] env[62521]: DEBUG nova.network.neutron [req-0735d331-1735-44b9-bb6f-eb6e819354d4 req-dfac5646-268f-4bf0-882c-643ee2071b51 service nova] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.479904] env[62521]: INFO nova.compute.manager [-] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Took 1.03 seconds to deallocate network for instance. [ 680.483423] env[62521]: DEBUG nova.compute.claims [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 680.483628] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.695851] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.019s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.696519] env[62521]: ERROR nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c84fc1ca-c5e3-40ba-a1d2-265501cf1467, please check neutron logs for more information. [ 680.696519] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] Traceback (most recent call last): [ 680.696519] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.696519] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] self.driver.spawn(context, instance, image_meta, [ 680.696519] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 680.696519] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.696519] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.696519] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] vm_ref = self.build_virtual_machine(instance, [ 680.696519] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.696519] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.696519] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] for vif in network_info: [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] return self._sync_wrapper(fn, *args, **kwargs) [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] self.wait() [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] self[:] = self._gt.wait() [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] return self._exit_event.wait() [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] result = hub.switch() [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.696894] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] return self.greenlet.switch() [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] result = function(*args, **kwargs) [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] return func(*args, **kwargs) [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] raise e [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] nwinfo = self.network_api.allocate_for_instance( [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] created_port_ids = self._update_ports_for_instance( [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] with excutils.save_and_reraise_exception(): [ 680.697289] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] self.force_reraise() [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] raise self.value [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] updated_port = self._update_port( [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] _ensure_no_port_binding_failure(port) [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] raise exception.PortBindingFailed(port_id=port['id']) [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] nova.exception.PortBindingFailed: Binding failed for port c84fc1ca-c5e3-40ba-a1d2-265501cf1467, please check neutron logs for more information. [ 680.697664] env[62521]: ERROR nova.compute.manager [instance: fe69caab-dd71-4b01-a121-2591111b193d] [ 680.698091] env[62521]: DEBUG nova.compute.utils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Binding failed for port c84fc1ca-c5e3-40ba-a1d2-265501cf1467, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.699111] env[62521]: DEBUG nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 680.701346] env[62521]: DEBUG nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Build of instance fe69caab-dd71-4b01-a121-2591111b193d was re-scheduled: Binding failed for port c84fc1ca-c5e3-40ba-a1d2-265501cf1467, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 680.701938] env[62521]: DEBUG nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 680.701979] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquiring lock "refresh_cache-fe69caab-dd71-4b01-a121-2591111b193d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.702124] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Acquired lock "refresh_cache-fe69caab-dd71-4b01-a121-2591111b193d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.702284] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.703228] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.839s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.732514] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 680.732772] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 680.732929] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 680.733198] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 680.733388] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 680.733575] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 680.733758] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 680.733921] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 680.734142] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 680.734320] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 680.734521] env[62521]: DEBUG nova.virt.hardware [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 680.735428] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b951d5c-1e07-4160-af49-0918592631af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.744896] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d00397a-cf0b-4f91-9cae-4a07ee26f628 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.885183] env[62521]: ERROR nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f66095d6-0bf8-4cfc-9921-8fda94e576ea, please check neutron logs for more information. [ 680.885183] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.885183] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.885183] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.885183] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.885183] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.885183] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.885183] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.885183] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.885183] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 680.885183] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.885183] env[62521]: ERROR nova.compute.manager raise self.value [ 680.885183] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.885183] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.885183] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.885183] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.885838] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.885838] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.885838] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f66095d6-0bf8-4cfc-9921-8fda94e576ea, please check neutron logs for more information. [ 680.885838] env[62521]: ERROR nova.compute.manager [ 680.885838] env[62521]: Traceback (most recent call last): [ 680.885838] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.885838] env[62521]: listener.cb(fileno) [ 680.885838] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.885838] env[62521]: result = function(*args, **kwargs) [ 680.885838] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.885838] env[62521]: return func(*args, **kwargs) [ 680.885838] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.885838] env[62521]: raise e [ 680.885838] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.885838] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 680.885838] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.885838] env[62521]: created_port_ids = self._update_ports_for_instance( [ 680.885838] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.885838] env[62521]: with excutils.save_and_reraise_exception(): [ 680.885838] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.885838] env[62521]: self.force_reraise() [ 680.885838] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.885838] env[62521]: raise self.value [ 680.885838] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.885838] env[62521]: updated_port = self._update_port( [ 680.885838] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.885838] env[62521]: _ensure_no_port_binding_failure(port) [ 680.885838] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.885838] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.886603] env[62521]: nova.exception.PortBindingFailed: Binding failed for port f66095d6-0bf8-4cfc-9921-8fda94e576ea, please check neutron logs for more information. [ 680.886603] env[62521]: Removing descriptor: 15 [ 680.886603] env[62521]: ERROR nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f66095d6-0bf8-4cfc-9921-8fda94e576ea, please check neutron logs for more information. [ 680.886603] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Traceback (most recent call last): [ 680.886603] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 680.886603] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] yield resources [ 680.886603] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.886603] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] self.driver.spawn(context, instance, image_meta, [ 680.886603] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 680.886603] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.886603] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.886603] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] vm_ref = self.build_virtual_machine(instance, [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] for vif in network_info: [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] return self._sync_wrapper(fn, *args, **kwargs) [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] self.wait() [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] self[:] = self._gt.wait() [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] return self._exit_event.wait() [ 680.886931] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] result = hub.switch() [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] return self.greenlet.switch() [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] result = function(*args, **kwargs) [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] return func(*args, **kwargs) [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] raise e [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] nwinfo = self.network_api.allocate_for_instance( [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.887406] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] created_port_ids = self._update_ports_for_instance( [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] with excutils.save_and_reraise_exception(): [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] self.force_reraise() [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] raise self.value [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] updated_port = self._update_port( [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] _ensure_no_port_binding_failure(port) [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.887767] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] raise exception.PortBindingFailed(port_id=port['id']) [ 680.888164] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] nova.exception.PortBindingFailed: Binding failed for port f66095d6-0bf8-4cfc-9921-8fda94e576ea, please check neutron logs for more information. [ 680.888164] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] [ 680.888164] env[62521]: INFO nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Terminating instance [ 680.888254] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Acquiring lock "refresh_cache-cd3d793a-aa56-4f07-a2fa-c56693548dfb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.888365] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Acquired lock "refresh_cache-cd3d793a-aa56-4f07-a2fa-c56693548dfb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.888530] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.943440] env[62521]: DEBUG oslo_concurrency.lockutils [req-0735d331-1735-44b9-bb6f-eb6e819354d4 req-dfac5646-268f-4bf0-882c-643ee2071b51 service nova] Releasing lock "refresh_cache-166786ee-4b89-4683-831b-9f3c1c57fb08" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.943703] env[62521]: DEBUG nova.compute.manager [req-0735d331-1735-44b9-bb6f-eb6e819354d4 req-dfac5646-268f-4bf0-882c-643ee2071b51 service nova] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Received event network-vif-deleted-fc7e5ec2-ec04-441b-86c3-97b6288e0367 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.231465] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.332236] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.405853] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.489225] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.600197] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56125400-e766-4c15-ba22-fd33bd1ad3ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.607750] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e2b457-2449-4475-a431-98fe330df80a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.636595] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3372bfed-c35c-4320-8da3-e0088b54b94b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.643284] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1246b72f-6d36-4479-8043-d26f1638ef07 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.657019] env[62521]: DEBUG nova.compute.provider_tree [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.834493] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Releasing lock "refresh_cache-fe69caab-dd71-4b01-a121-2591111b193d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.834805] env[62521]: DEBUG nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.834994] env[62521]: DEBUG nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.835181] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.851525] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.903686] env[62521]: DEBUG nova.compute.manager [req-a62475a7-f851-4fb1-b14f-0876a768b52c req-b1c0162c-5389-42ad-a132-e33b73d42b21 service nova] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Received event network-changed-f66095d6-0bf8-4cfc-9921-8fda94e576ea {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.903999] env[62521]: DEBUG nova.compute.manager [req-a62475a7-f851-4fb1-b14f-0876a768b52c req-b1c0162c-5389-42ad-a132-e33b73d42b21 service nova] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Refreshing instance network info cache due to event network-changed-f66095d6-0bf8-4cfc-9921-8fda94e576ea. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 681.904229] env[62521]: DEBUG oslo_concurrency.lockutils [req-a62475a7-f851-4fb1-b14f-0876a768b52c req-b1c0162c-5389-42ad-a132-e33b73d42b21 service nova] Acquiring lock "refresh_cache-cd3d793a-aa56-4f07-a2fa-c56693548dfb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.994450] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Releasing lock "refresh_cache-cd3d793a-aa56-4f07-a2fa-c56693548dfb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.994689] env[62521]: DEBUG nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 681.994879] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 681.995200] env[62521]: DEBUG oslo_concurrency.lockutils [req-a62475a7-f851-4fb1-b14f-0876a768b52c req-b1c0162c-5389-42ad-a132-e33b73d42b21 service nova] Acquired lock "refresh_cache-cd3d793a-aa56-4f07-a2fa-c56693548dfb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.995372] env[62521]: DEBUG nova.network.neutron [req-a62475a7-f851-4fb1-b14f-0876a768b52c req-b1c0162c-5389-42ad-a132-e33b73d42b21 service nova] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Refreshing network info cache for port f66095d6-0bf8-4cfc-9921-8fda94e576ea {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 681.996392] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3134d76d-6a66-45e2-bcff-01bf217778e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.006607] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd34a47c-070e-488b-8125-31b09dcde9c0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.028750] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cd3d793a-aa56-4f07-a2fa-c56693548dfb could not be found. [ 682.028968] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 682.029161] env[62521]: INFO nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 682.029399] env[62521]: DEBUG oslo.service.loopingcall [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 682.029614] env[62521]: DEBUG nova.compute.manager [-] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.029707] env[62521]: DEBUG nova.network.neutron [-] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 682.056311] env[62521]: DEBUG nova.network.neutron [-] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.159649] env[62521]: DEBUG nova.scheduler.client.report [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.353700] env[62521]: DEBUG nova.network.neutron [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.523458] env[62521]: DEBUG nova.network.neutron [req-a62475a7-f851-4fb1-b14f-0876a768b52c req-b1c0162c-5389-42ad-a132-e33b73d42b21 service nova] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.558726] env[62521]: DEBUG nova.network.neutron [-] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.619888] env[62521]: DEBUG nova.network.neutron [req-a62475a7-f851-4fb1-b14f-0876a768b52c req-b1c0162c-5389-42ad-a132-e33b73d42b21 service nova] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.665757] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.666404] env[62521]: ERROR nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 47f7f57d-d2dc-4ce6-8259-b7a863189fa7, please check neutron logs for more information. [ 682.666404] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Traceback (most recent call last): [ 682.666404] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.666404] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] self.driver.spawn(context, instance, image_meta, [ 682.666404] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 682.666404] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.666404] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.666404] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] vm_ref = self.build_virtual_machine(instance, [ 682.666404] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.666404] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.666404] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] for vif in network_info: [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] return self._sync_wrapper(fn, *args, **kwargs) [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] self.wait() [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] self[:] = self._gt.wait() [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] return self._exit_event.wait() [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] result = hub.switch() [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 682.666765] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] return self.greenlet.switch() [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] result = function(*args, **kwargs) [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] return func(*args, **kwargs) [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] raise e [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] nwinfo = self.network_api.allocate_for_instance( [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] created_port_ids = self._update_ports_for_instance( [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] with excutils.save_and_reraise_exception(): [ 682.667171] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] self.force_reraise() [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] raise self.value [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] updated_port = self._update_port( [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] _ensure_no_port_binding_failure(port) [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] raise exception.PortBindingFailed(port_id=port['id']) [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] nova.exception.PortBindingFailed: Binding failed for port 47f7f57d-d2dc-4ce6-8259-b7a863189fa7, please check neutron logs for more information. [ 682.667545] env[62521]: ERROR nova.compute.manager [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] [ 682.667865] env[62521]: DEBUG nova.compute.utils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Binding failed for port 47f7f57d-d2dc-4ce6-8259-b7a863189fa7, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.668208] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.507s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.669744] env[62521]: INFO nova.compute.claims [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.672349] env[62521]: DEBUG nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Build of instance edd89a57-f026-4a3e-8b6b-111cc2321d9d was re-scheduled: Binding failed for port 47f7f57d-d2dc-4ce6-8259-b7a863189fa7, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 682.672773] env[62521]: DEBUG nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 682.672999] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Acquiring lock "refresh_cache-edd89a57-f026-4a3e-8b6b-111cc2321d9d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.673163] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Acquired lock "refresh_cache-edd89a57-f026-4a3e-8b6b-111cc2321d9d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.673323] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.856891] env[62521]: INFO nova.compute.manager [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] [instance: fe69caab-dd71-4b01-a121-2591111b193d] Took 1.02 seconds to deallocate network for instance. [ 683.061706] env[62521]: INFO nova.compute.manager [-] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Took 1.03 seconds to deallocate network for instance. [ 683.066977] env[62521]: DEBUG nova.compute.claims [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 683.067198] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.121945] env[62521]: DEBUG oslo_concurrency.lockutils [req-a62475a7-f851-4fb1-b14f-0876a768b52c req-b1c0162c-5389-42ad-a132-e33b73d42b21 service nova] Releasing lock "refresh_cache-cd3d793a-aa56-4f07-a2fa-c56693548dfb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.122240] env[62521]: DEBUG nova.compute.manager [req-a62475a7-f851-4fb1-b14f-0876a768b52c req-b1c0162c-5389-42ad-a132-e33b73d42b21 service nova] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Received event network-vif-deleted-f66095d6-0bf8-4cfc-9921-8fda94e576ea {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 683.193044] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.307075] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.810408] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Releasing lock "refresh_cache-edd89a57-f026-4a3e-8b6b-111cc2321d9d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.810677] env[62521]: DEBUG nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 683.810810] env[62521]: DEBUG nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.810978] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.833118] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.881608] env[62521]: INFO nova.scheduler.client.report [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Deleted allocations for instance fe69caab-dd71-4b01-a121-2591111b193d [ 684.037282] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625694ad-3202-4936-be10-e681caaf3ad3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.045567] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e789bce-59cd-4c03-be9e-ced97a102ebb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.074631] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4029c0-8745-47dc-beae-1b61f8a6e18f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.082292] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b670b1c8-78d0-4fdc-a438-a65e020fabed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.095433] env[62521]: DEBUG nova.compute.provider_tree [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.336377] env[62521]: DEBUG nova.network.neutron [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.393737] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25bfabb0-7405-4e48-91ce-826125d43550 tempest-ServersAdminTestJSON-1576042830 tempest-ServersAdminTestJSON-1576042830-project-member] Lock "fe69caab-dd71-4b01-a121-2591111b193d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.946s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.598781] env[62521]: DEBUG nova.scheduler.client.report [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.839041] env[62521]: INFO nova.compute.manager [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] [instance: edd89a57-f026-4a3e-8b6b-111cc2321d9d] Took 1.03 seconds to deallocate network for instance. [ 684.896847] env[62521]: DEBUG nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.103710] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.104284] env[62521]: DEBUG nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 685.107310] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.641s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.433821] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.611116] env[62521]: DEBUG nova.compute.utils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.617198] env[62521]: DEBUG nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 685.617301] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 685.662615] env[62521]: DEBUG nova.policy [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '016f2b75120648999bdb139b28f60900', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e84f7147da104173b0807ea29b94cf41', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 685.874768] env[62521]: INFO nova.scheduler.client.report [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Deleted allocations for instance edd89a57-f026-4a3e-8b6b-111cc2321d9d [ 686.118614] env[62521]: DEBUG nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 686.163803] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 104b64ff-a042-4bee-a5fd-bbbdf288f63f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 686.163803] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance e0aabeb7-7d4c-4358-8f6c-e423a0baea3b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 686.163803] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 215472ea-402c-4eb5-8ad3-b7ef756dfdc5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 686.163803] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance f7611f03-73bb-4c0c-b787-d3c7f077f5bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 686.163932] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance e1ead278-d904-41cc-bdfe-65174e8d6816 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 686.163932] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 166786ee-4b89-4683-831b-9f3c1c57fb08 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 686.163932] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance cd3d793a-aa56-4f07-a2fa-c56693548dfb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 686.163932] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance a4965864-0127-4842-9478-76e369e2ecbe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 686.186516] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Successfully created port: 67d58d1f-300c-497f-af25-825bc30fae23 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.385986] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8dd2331f-0c52-4729-a2dc-48b937406a73 tempest-ServerTagsTestJSON-1795941190 tempest-ServerTagsTestJSON-1795941190-project-member] Lock "edd89a57-f026-4a3e-8b6b-111cc2321d9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.173s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.668154] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance cf71efc9-cfa1-4856-a733-5434bf4465b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.890397] env[62521]: DEBUG nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.130882] env[62521]: DEBUG nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 687.157261] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.157506] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.157740] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.157841] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.157989] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.158148] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.158349] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.158507] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.158749] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.158817] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.158981] env[62521]: DEBUG nova.virt.hardware [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.159869] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0192a36-56ee-41a3-adf2-2c2b238ecc2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.167945] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d10c2c-537c-47a0-86b8-c5d494b8a274 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.173390] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance d7575b46-ecf9-4154-92b1-86119e727dda has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.423280] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.563656] env[62521]: DEBUG nova.compute.manager [req-a4077598-0ccd-42bc-ba4b-d082865a3cfe req-60ff5d0b-bb5f-47b3-9a4e-f0586c1f673d service nova] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Received event network-changed-67d58d1f-300c-497f-af25-825bc30fae23 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 687.563860] env[62521]: DEBUG nova.compute.manager [req-a4077598-0ccd-42bc-ba4b-d082865a3cfe req-60ff5d0b-bb5f-47b3-9a4e-f0586c1f673d service nova] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Refreshing instance network info cache due to event network-changed-67d58d1f-300c-497f-af25-825bc30fae23. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 687.564105] env[62521]: DEBUG oslo_concurrency.lockutils [req-a4077598-0ccd-42bc-ba4b-d082865a3cfe req-60ff5d0b-bb5f-47b3-9a4e-f0586c1f673d service nova] Acquiring lock "refresh_cache-a4965864-0127-4842-9478-76e369e2ecbe" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.564267] env[62521]: DEBUG oslo_concurrency.lockutils [req-a4077598-0ccd-42bc-ba4b-d082865a3cfe req-60ff5d0b-bb5f-47b3-9a4e-f0586c1f673d service nova] Acquired lock "refresh_cache-a4965864-0127-4842-9478-76e369e2ecbe" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.564457] env[62521]: DEBUG nova.network.neutron [req-a4077598-0ccd-42bc-ba4b-d082865a3cfe req-60ff5d0b-bb5f-47b3-9a4e-f0586c1f673d service nova] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Refreshing network info cache for port 67d58d1f-300c-497f-af25-825bc30fae23 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 687.677164] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 0d72d988-f96b-428c-89bf-a96572f71db5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.933875] env[62521]: ERROR nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 67d58d1f-300c-497f-af25-825bc30fae23, please check neutron logs for more information. [ 687.933875] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 687.933875] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.933875] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 687.933875] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.933875] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 687.933875] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.933875] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 687.933875] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.933875] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 687.933875] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.933875] env[62521]: ERROR nova.compute.manager raise self.value [ 687.933875] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.933875] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 687.933875] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.933875] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 687.934794] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.934794] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 687.934794] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 67d58d1f-300c-497f-af25-825bc30fae23, please check neutron logs for more information. [ 687.934794] env[62521]: ERROR nova.compute.manager [ 687.934794] env[62521]: Traceback (most recent call last): [ 687.934794] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 687.934794] env[62521]: listener.cb(fileno) [ 687.934794] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.934794] env[62521]: result = function(*args, **kwargs) [ 687.934794] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.934794] env[62521]: return func(*args, **kwargs) [ 687.934794] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.934794] env[62521]: raise e [ 687.934794] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.934794] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 687.934794] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.934794] env[62521]: created_port_ids = self._update_ports_for_instance( [ 687.934794] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.934794] env[62521]: with excutils.save_and_reraise_exception(): [ 687.934794] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.934794] env[62521]: self.force_reraise() [ 687.934794] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.934794] env[62521]: raise self.value [ 687.934794] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.934794] env[62521]: updated_port = self._update_port( [ 687.934794] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.934794] env[62521]: _ensure_no_port_binding_failure(port) [ 687.934794] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.934794] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 687.935555] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 67d58d1f-300c-497f-af25-825bc30fae23, please check neutron logs for more information. [ 687.935555] env[62521]: Removing descriptor: 15 [ 687.935555] env[62521]: ERROR nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 67d58d1f-300c-497f-af25-825bc30fae23, please check neutron logs for more information. [ 687.935555] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] Traceback (most recent call last): [ 687.935555] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 687.935555] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] yield resources [ 687.935555] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.935555] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] self.driver.spawn(context, instance, image_meta, [ 687.935555] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 687.935555] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.935555] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.935555] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] vm_ref = self.build_virtual_machine(instance, [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] for vif in network_info: [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] return self._sync_wrapper(fn, *args, **kwargs) [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] self.wait() [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] self[:] = self._gt.wait() [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] return self._exit_event.wait() [ 687.935873] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] result = hub.switch() [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] return self.greenlet.switch() [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] result = function(*args, **kwargs) [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] return func(*args, **kwargs) [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] raise e [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] nwinfo = self.network_api.allocate_for_instance( [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.936193] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] created_port_ids = self._update_ports_for_instance( [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] with excutils.save_and_reraise_exception(): [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] self.force_reraise() [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] raise self.value [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] updated_port = self._update_port( [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] _ensure_no_port_binding_failure(port) [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.936509] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] raise exception.PortBindingFailed(port_id=port['id']) [ 687.936846] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] nova.exception.PortBindingFailed: Binding failed for port 67d58d1f-300c-497f-af25-825bc30fae23, please check neutron logs for more information. [ 687.936846] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] [ 687.936846] env[62521]: INFO nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Terminating instance [ 687.940071] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Acquiring lock "refresh_cache-a4965864-0127-4842-9478-76e369e2ecbe" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.086165] env[62521]: DEBUG nova.network.neutron [req-a4077598-0ccd-42bc-ba4b-d082865a3cfe req-60ff5d0b-bb5f-47b3-9a4e-f0586c1f673d service nova] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.185908] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 5fa1c73e-3074-4fe0-8b4d-926bf02f1453 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.267389] env[62521]: DEBUG nova.network.neutron [req-a4077598-0ccd-42bc-ba4b-d082865a3cfe req-60ff5d0b-bb5f-47b3-9a4e-f0586c1f673d service nova] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.692651] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 884d2f4f-c0a2-4d08-962e-d6347dc9f678 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.770234] env[62521]: DEBUG oslo_concurrency.lockutils [req-a4077598-0ccd-42bc-ba4b-d082865a3cfe req-60ff5d0b-bb5f-47b3-9a4e-f0586c1f673d service nova] Releasing lock "refresh_cache-a4965864-0127-4842-9478-76e369e2ecbe" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.773037] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Acquired lock "refresh_cache-a4965864-0127-4842-9478-76e369e2ecbe" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.773037] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.196210] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 79b55308-ff33-4b2a-bca8-30a1c5cf5916 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.295161] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.466994] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.701821] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 5adbe047-7055-428f-b72b-67023c68cc04 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.712676] env[62521]: DEBUG nova.compute.manager [req-16f9364e-7980-4d66-a8ed-a0cba44f1069 req-062eb2d8-25c4-46a5-b227-97c95500c131 service nova] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Received event network-vif-deleted-67d58d1f-300c-497f-af25-825bc30fae23 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 689.972327] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Releasing lock "refresh_cache-a4965864-0127-4842-9478-76e369e2ecbe" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.973610] env[62521]: DEBUG nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 689.973610] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 689.973610] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50a4e047-7d15-4b2d-bb41-d2be1151d49a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.984359] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39741442-fcd9-4ac5-b61d-a7fbb11cc8fc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.007512] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a4965864-0127-4842-9478-76e369e2ecbe could not be found. [ 690.007571] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.007732] env[62521]: INFO nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Took 0.03 seconds to destroy the instance on the hypervisor. [ 690.007983] env[62521]: DEBUG oslo.service.loopingcall [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.008218] env[62521]: DEBUG nova.compute.manager [-] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.008314] env[62521]: DEBUG nova.network.neutron [-] [instance: a4965864-0127-4842-9478-76e369e2ecbe] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.025720] env[62521]: DEBUG nova.network.neutron [-] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.205533] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 72e3b843-c54a-4de7-ace9-ee25c5df14bd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.529571] env[62521]: DEBUG nova.network.neutron [-] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.710445] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 86727bdc-342a-429c-85ce-adec8823b799 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.033186] env[62521]: INFO nova.compute.manager [-] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Took 1.02 seconds to deallocate network for instance. [ 691.035738] env[62521]: DEBUG nova.compute.claims [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 691.035944] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.217099] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 277ecbd7-e770-4ea0-ae2e-b49598428d0f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.721412] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.153502] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "b3e33bab-636f-42df-ac6c-667464312c11" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.153634] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "b3e33bab-636f-42df-ac6c-667464312c11" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.226946] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 32c0268b-eee6-4e45-b9c3-077ec81bd314 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.732416] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 63abbe0a-af36-4679-8cf1-87f9e9abf7b9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.235653] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance f6a27e36-1728-4056-99a6-ed2114df64ce has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.409846] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.410103] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.738557] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 6be7d826-e19b-431e-b714-25e646e6b7ca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.242027] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 590b2441-e627-463b-9722-d0fcb8fcd83a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.746056] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 062f311c-a783-41cf-8d83-b3a8d4df14f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.249066] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 868c7798-8d0f-4c31-88dc-5007454796f1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.752540] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance de583c0b-d147-4f54-bc90-49bdd71c56aa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.255969] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.759102] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 50aef23e-9429-4499-9fc1-c31b97e23029 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.264099] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.765788] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance e8cabcdc-d728-434d-a40d-fe751db086c3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.766108] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 697.766264] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 698.129828] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1b37a7-5628-47a5-96c8-190cc0080549 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.137537] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c0cb86-62ac-4f1e-86f5-11d60290c76c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.167213] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e8191e-37d4-452f-9150-098c15d9ae38 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.174104] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c61dbee-4afd-4a90-85d2-e33c72d1afa4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.186652] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.690467] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.197637] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 699.197637] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.088s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.197637] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.656s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.049891] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51adbfe-12b8-4e5b-a058-9ad7c9dbfe1f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.057497] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8045652b-5f9b-46c7-81fb-da82b2b14548 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.086791] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf9c8a7-8ad2-4207-9013-869da6dab26e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.093674] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9f689b-0efd-472a-8463-f309fd999213 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.107349] env[62521]: DEBUG nova.compute.provider_tree [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.610541] env[62521]: DEBUG nova.scheduler.client.report [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.115338] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.919s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.116045] env[62521]: ERROR nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ef49cd51-3ad9-4500-a7fb-42b6cb8e8916, please check neutron logs for more information. [ 701.116045] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Traceback (most recent call last): [ 701.116045] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.116045] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] self.driver.spawn(context, instance, image_meta, [ 701.116045] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 701.116045] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.116045] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.116045] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] vm_ref = self.build_virtual_machine(instance, [ 701.116045] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.116045] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.116045] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] for vif in network_info: [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] return self._sync_wrapper(fn, *args, **kwargs) [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] self.wait() [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] self[:] = self._gt.wait() [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] return self._exit_event.wait() [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] result = hub.switch() [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.116344] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] return self.greenlet.switch() [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] result = function(*args, **kwargs) [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] return func(*args, **kwargs) [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] raise e [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] nwinfo = self.network_api.allocate_for_instance( [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] created_port_ids = self._update_ports_for_instance( [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] with excutils.save_and_reraise_exception(): [ 701.116665] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] self.force_reraise() [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] raise self.value [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] updated_port = self._update_port( [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] _ensure_no_port_binding_failure(port) [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] raise exception.PortBindingFailed(port_id=port['id']) [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] nova.exception.PortBindingFailed: Binding failed for port ef49cd51-3ad9-4500-a7fb-42b6cb8e8916, please check neutron logs for more information. [ 701.116990] env[62521]: ERROR nova.compute.manager [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] [ 701.117544] env[62521]: DEBUG nova.compute.utils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Binding failed for port ef49cd51-3ad9-4500-a7fb-42b6cb8e8916, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.118033] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.780s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.120856] env[62521]: DEBUG nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Build of instance 104b64ff-a042-4bee-a5fd-bbbdf288f63f was re-scheduled: Binding failed for port ef49cd51-3ad9-4500-a7fb-42b6cb8e8916, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 701.121301] env[62521]: DEBUG nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 701.121522] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquiring lock "refresh_cache-104b64ff-a042-4bee-a5fd-bbbdf288f63f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.121665] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Acquired lock "refresh_cache-104b64ff-a042-4bee-a5fd-bbbdf288f63f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.121825] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.654199] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.702107] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.978855] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da885ae-810e-4504-ac66-a781f8040245 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.986699] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04eac533-c9da-4e4c-8375-e667e34bf0a2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.015531] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03422548-eeef-4ec9-a0ef-856d3b82ba09 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.022857] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b70f58-ab3d-460d-985a-62eb96390902 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.036736] env[62521]: DEBUG nova.compute.provider_tree [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.206280] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Releasing lock "refresh_cache-104b64ff-a042-4bee-a5fd-bbbdf288f63f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.206563] env[62521]: DEBUG nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 702.206703] env[62521]: DEBUG nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.206871] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.219993] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.539672] env[62521]: DEBUG nova.scheduler.client.report [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.722575] env[62521]: DEBUG nova.network.neutron [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.044076] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.926s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.044767] env[62521]: ERROR nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6, please check neutron logs for more information. [ 703.044767] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Traceback (most recent call last): [ 703.044767] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.044767] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] self.driver.spawn(context, instance, image_meta, [ 703.044767] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 703.044767] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.044767] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.044767] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] vm_ref = self.build_virtual_machine(instance, [ 703.044767] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.044767] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.044767] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] for vif in network_info: [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] return self._sync_wrapper(fn, *args, **kwargs) [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] self.wait() [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] self[:] = self._gt.wait() [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] return self._exit_event.wait() [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] result = hub.switch() [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 703.045298] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] return self.greenlet.switch() [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] result = function(*args, **kwargs) [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] return func(*args, **kwargs) [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] raise e [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] nwinfo = self.network_api.allocate_for_instance( [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] created_port_ids = self._update_ports_for_instance( [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] with excutils.save_and_reraise_exception(): [ 703.045711] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] self.force_reraise() [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] raise self.value [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] updated_port = self._update_port( [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] _ensure_no_port_binding_failure(port) [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] raise exception.PortBindingFailed(port_id=port['id']) [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] nova.exception.PortBindingFailed: Binding failed for port 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6, please check neutron logs for more information. [ 703.046158] env[62521]: ERROR nova.compute.manager [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] [ 703.046452] env[62521]: DEBUG nova.compute.utils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Binding failed for port 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.046749] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.353s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.049903] env[62521]: DEBUG nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Build of instance e0aabeb7-7d4c-4358-8f6c-e423a0baea3b was re-scheduled: Binding failed for port 3f54e9d5-40a8-4d3f-ab5f-0292a78229a6, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 703.050269] env[62521]: DEBUG nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 703.050496] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Acquiring lock "refresh_cache-e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.050642] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Acquired lock "refresh_cache-e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.050901] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.226100] env[62521]: INFO nova.compute.manager [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] [instance: 104b64ff-a042-4bee-a5fd-bbbdf288f63f] Took 1.02 seconds to deallocate network for instance. [ 703.571243] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.735958] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.961210] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458cf843-2468-40a3-8cb9-45e4946141b7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.968789] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cf7760-8727-4ba4-b719-c7657d314339 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.998894] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b690429-8cfc-4d67-9572-e6306a0c1a54 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.006066] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989851ee-dffc-45c3-ad5f-479bdc2dc238 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.019060] env[62521]: DEBUG nova.compute.provider_tree [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.240697] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Releasing lock "refresh_cache-e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.240996] env[62521]: DEBUG nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 704.241116] env[62521]: DEBUG nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.241264] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.260131] env[62521]: INFO nova.scheduler.client.report [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Deleted allocations for instance 104b64ff-a042-4bee-a5fd-bbbdf288f63f [ 704.265573] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.523015] env[62521]: DEBUG nova.scheduler.client.report [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.768553] env[62521]: DEBUG nova.network.neutron [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.769899] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d590e325-2efe-4170-90d3-3348c28eaaf5 tempest-DeleteServersAdminTestJSON-1875485469 tempest-DeleteServersAdminTestJSON-1875485469-project-member] Lock "104b64ff-a042-4bee-a5fd-bbbdf288f63f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.962s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.027383] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.028074] env[62521]: ERROR nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6cee9131-a87f-4722-88b3-00a7ba0882ae, please check neutron logs for more information. [ 705.028074] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Traceback (most recent call last): [ 705.028074] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.028074] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] self.driver.spawn(context, instance, image_meta, [ 705.028074] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 705.028074] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.028074] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.028074] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] vm_ref = self.build_virtual_machine(instance, [ 705.028074] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.028074] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.028074] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] for vif in network_info: [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] return self._sync_wrapper(fn, *args, **kwargs) [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] self.wait() [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] self[:] = self._gt.wait() [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] return self._exit_event.wait() [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] result = hub.switch() [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.028467] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] return self.greenlet.switch() [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] result = function(*args, **kwargs) [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] return func(*args, **kwargs) [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] raise e [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] nwinfo = self.network_api.allocate_for_instance( [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] created_port_ids = self._update_ports_for_instance( [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] with excutils.save_and_reraise_exception(): [ 705.028898] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] self.force_reraise() [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] raise self.value [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] updated_port = self._update_port( [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] _ensure_no_port_binding_failure(port) [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] raise exception.PortBindingFailed(port_id=port['id']) [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] nova.exception.PortBindingFailed: Binding failed for port 6cee9131-a87f-4722-88b3-00a7ba0882ae, please check neutron logs for more information. [ 705.029225] env[62521]: ERROR nova.compute.manager [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] [ 705.029867] env[62521]: DEBUG nova.compute.utils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Binding failed for port 6cee9131-a87f-4722-88b3-00a7ba0882ae, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 705.030106] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.647s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.033290] env[62521]: DEBUG nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Build of instance 215472ea-402c-4eb5-8ad3-b7ef756dfdc5 was re-scheduled: Binding failed for port 6cee9131-a87f-4722-88b3-00a7ba0882ae, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 705.033728] env[62521]: DEBUG nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 705.033953] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "refresh_cache-215472ea-402c-4eb5-8ad3-b7ef756dfdc5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.034112] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "refresh_cache-215472ea-402c-4eb5-8ad3-b7ef756dfdc5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.034268] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.274011] env[62521]: INFO nova.compute.manager [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] [instance: e0aabeb7-7d4c-4358-8f6c-e423a0baea3b] Took 1.03 seconds to deallocate network for instance. [ 705.276600] env[62521]: DEBUG nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 705.558598] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.692701] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.801965] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.975019] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abdb2245-eca9-4a90-aa7d-3516af02dc34 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.982097] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98768f0c-ae92-45f8-96cd-1b4e60eeaf4e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.018358] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b689223-27a2-4698-b216-452a56dab224 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.030014] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c098de4-6083-4b05-9994-f5a4665aa349 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.040451] env[62521]: DEBUG nova.compute.provider_tree [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.196626] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "refresh_cache-215472ea-402c-4eb5-8ad3-b7ef756dfdc5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.197926] env[62521]: DEBUG nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 706.197926] env[62521]: DEBUG nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.197926] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.229171] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.309091] env[62521]: INFO nova.scheduler.client.report [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Deleted allocations for instance e0aabeb7-7d4c-4358-8f6c-e423a0baea3b [ 706.544320] env[62521]: DEBUG nova.scheduler.client.report [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.731317] env[62521]: DEBUG nova.network.neutron [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.822057] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a6195d60-b6fb-40ef-a5e1-95acb7880525 tempest-ImagesNegativeTestJSON-1874109051 tempest-ImagesNegativeTestJSON-1874109051-project-member] Lock "e0aabeb7-7d4c-4358-8f6c-e423a0baea3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.079s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.049978] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.020s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.050659] env[62521]: ERROR nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c9f4099c-087b-4d78-b799-e54699bcfcd9, please check neutron logs for more information. [ 707.050659] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Traceback (most recent call last): [ 707.050659] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.050659] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] self.driver.spawn(context, instance, image_meta, [ 707.050659] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 707.050659] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.050659] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.050659] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] vm_ref = self.build_virtual_machine(instance, [ 707.050659] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.050659] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.050659] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] for vif in network_info: [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] return self._sync_wrapper(fn, *args, **kwargs) [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] self.wait() [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] self[:] = self._gt.wait() [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] return self._exit_event.wait() [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] result = hub.switch() [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 707.050984] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] return self.greenlet.switch() [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] result = function(*args, **kwargs) [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] return func(*args, **kwargs) [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] raise e [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] nwinfo = self.network_api.allocate_for_instance( [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] created_port_ids = self._update_ports_for_instance( [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] with excutils.save_and_reraise_exception(): [ 707.051355] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] self.force_reraise() [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] raise self.value [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] updated_port = self._update_port( [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] _ensure_no_port_binding_failure(port) [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] raise exception.PortBindingFailed(port_id=port['id']) [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] nova.exception.PortBindingFailed: Binding failed for port c9f4099c-087b-4d78-b799-e54699bcfcd9, please check neutron logs for more information. [ 707.051682] env[62521]: ERROR nova.compute.manager [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] [ 707.051953] env[62521]: DEBUG nova.compute.utils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Binding failed for port c9f4099c-087b-4d78-b799-e54699bcfcd9, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.053141] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.923s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.055951] env[62521]: DEBUG nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Build of instance f7611f03-73bb-4c0c-b787-d3c7f077f5bf was re-scheduled: Binding failed for port c9f4099c-087b-4d78-b799-e54699bcfcd9, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 707.056439] env[62521]: DEBUG nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 707.056730] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-f7611f03-73bb-4c0c-b787-d3c7f077f5bf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.056794] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-f7611f03-73bb-4c0c-b787-d3c7f077f5bf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.056989] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.237028] env[62521]: INFO nova.compute.manager [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 215472ea-402c-4eb5-8ad3-b7ef756dfdc5] Took 1.04 seconds to deallocate network for instance. [ 707.325532] env[62521]: DEBUG nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 707.587168] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.707497] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.842666] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.979833] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb8c1f6-6bf4-4ddf-9c06-21ce66392d28 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.987299] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1801bd35-40c7-43b1-8712-a38d2e20db85 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.018438] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568ef155-8142-49bc-931f-307383dca096 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.025847] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f47c6b1-eecc-406b-ba2c-89f5842c29b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.038878] env[62521]: DEBUG nova.compute.provider_tree [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.072121] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "d5f6492d-ca1b-4e74-b792-b04d55c33660" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.072283] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.214765] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-f7611f03-73bb-4c0c-b787-d3c7f077f5bf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.215020] env[62521]: DEBUG nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 708.215204] env[62521]: DEBUG nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.215376] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.235431] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.266547] env[62521]: INFO nova.scheduler.client.report [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleted allocations for instance 215472ea-402c-4eb5-8ad3-b7ef756dfdc5 [ 708.542744] env[62521]: DEBUG nova.scheduler.client.report [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.738281] env[62521]: DEBUG nova.network.neutron [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.774473] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e758cd2b-3b2c-444e-b17b-f3ef4af218ef tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "215472ea-402c-4eb5-8ad3-b7ef756dfdc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.647s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.050027] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.997s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.051916] env[62521]: ERROR nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3fce0f5c-936f-4ca0-94ca-d2cb13d63094, please check neutron logs for more information. [ 709.051916] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Traceback (most recent call last): [ 709.051916] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 709.051916] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] self.driver.spawn(context, instance, image_meta, [ 709.051916] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 709.051916] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.051916] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.051916] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] vm_ref = self.build_virtual_machine(instance, [ 709.051916] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.051916] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.051916] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] for vif in network_info: [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] return self._sync_wrapper(fn, *args, **kwargs) [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] self.wait() [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] self[:] = self._gt.wait() [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] return self._exit_event.wait() [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] current.throw(*self._exc) [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.052879] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] result = function(*args, **kwargs) [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] return func(*args, **kwargs) [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] raise e [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] nwinfo = self.network_api.allocate_for_instance( [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] created_port_ids = self._update_ports_for_instance( [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] with excutils.save_and_reraise_exception(): [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] self.force_reraise() [ 709.053525] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.054137] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] raise self.value [ 709.054137] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.054137] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] updated_port = self._update_port( [ 709.054137] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.054137] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] _ensure_no_port_binding_failure(port) [ 709.054137] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.054137] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] raise exception.PortBindingFailed(port_id=port['id']) [ 709.054137] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] nova.exception.PortBindingFailed: Binding failed for port 3fce0f5c-936f-4ca0-94ca-d2cb13d63094, please check neutron logs for more information. [ 709.054137] env[62521]: ERROR nova.compute.manager [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] [ 709.054137] env[62521]: DEBUG nova.compute.utils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Binding failed for port 3fce0f5c-936f-4ca0-94ca-d2cb13d63094, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 709.054617] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.570s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.060645] env[62521]: DEBUG nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Build of instance e1ead278-d904-41cc-bdfe-65174e8d6816 was re-scheduled: Binding failed for port 3fce0f5c-936f-4ca0-94ca-d2cb13d63094, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 709.060645] env[62521]: DEBUG nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 709.060645] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Acquiring lock "refresh_cache-e1ead278-d904-41cc-bdfe-65174e8d6816" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.060645] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Acquired lock "refresh_cache-e1ead278-d904-41cc-bdfe-65174e8d6816" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.060895] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 709.240619] env[62521]: INFO nova.compute.manager [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: f7611f03-73bb-4c0c-b787-d3c7f077f5bf] Took 1.03 seconds to deallocate network for instance. [ 709.280133] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 709.580975] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.680757] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.795417] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.952430] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867acc67-8cef-4d28-a68a-1464fad48377 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.960735] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed169ae1-b2ba-41c0-9707-f912c1285b1e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.991978] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275c9ec1-5c6e-4b8c-9511-eeb17d8c61d4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.000024] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e5533f-e9d6-4633-a322-1187ea0f2da9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.013720] env[62521]: DEBUG nova.compute.provider_tree [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.072405] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "4dcd1c6c-8726-42db-997a-e78ba1293310" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.072644] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "4dcd1c6c-8726-42db-997a-e78ba1293310" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.186136] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Releasing lock "refresh_cache-e1ead278-d904-41cc-bdfe-65174e8d6816" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.186136] env[62521]: DEBUG nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 710.186136] env[62521]: DEBUG nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.186136] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.213648] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.267078] env[62521]: INFO nova.scheduler.client.report [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted allocations for instance f7611f03-73bb-4c0c-b787-d3c7f077f5bf [ 710.516900] env[62521]: DEBUG nova.scheduler.client.report [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.718056] env[62521]: DEBUG nova.network.neutron [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.775110] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6ba77609-e654-4472-b579-df920c23c409 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "f7611f03-73bb-4c0c-b787-d3c7f077f5bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.993s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.022148] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.968s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.022823] env[62521]: ERROR nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fc7e5ec2-ec04-441b-86c3-97b6288e0367, please check neutron logs for more information. [ 711.022823] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Traceback (most recent call last): [ 711.022823] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.022823] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] self.driver.spawn(context, instance, image_meta, [ 711.022823] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 711.022823] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.022823] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.022823] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] vm_ref = self.build_virtual_machine(instance, [ 711.022823] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.022823] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.022823] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] for vif in network_info: [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] return self._sync_wrapper(fn, *args, **kwargs) [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] self.wait() [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] self[:] = self._gt.wait() [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] return self._exit_event.wait() [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] result = hub.switch() [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 711.023106] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] return self.greenlet.switch() [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] result = function(*args, **kwargs) [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] return func(*args, **kwargs) [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] raise e [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] nwinfo = self.network_api.allocate_for_instance( [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] created_port_ids = self._update_ports_for_instance( [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] with excutils.save_and_reraise_exception(): [ 711.023427] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] self.force_reraise() [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] raise self.value [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] updated_port = self._update_port( [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] _ensure_no_port_binding_failure(port) [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] raise exception.PortBindingFailed(port_id=port['id']) [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] nova.exception.PortBindingFailed: Binding failed for port fc7e5ec2-ec04-441b-86c3-97b6288e0367, please check neutron logs for more information. [ 711.023747] env[62521]: ERROR nova.compute.manager [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] [ 711.024028] env[62521]: DEBUG nova.compute.utils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Binding failed for port fc7e5ec2-ec04-441b-86c3-97b6288e0367, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 711.027512] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.959s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.031046] env[62521]: DEBUG nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Build of instance 166786ee-4b89-4683-831b-9f3c1c57fb08 was re-scheduled: Binding failed for port fc7e5ec2-ec04-441b-86c3-97b6288e0367, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 711.031721] env[62521]: DEBUG nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 711.031721] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Acquiring lock "refresh_cache-166786ee-4b89-4683-831b-9f3c1c57fb08" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.031850] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Acquired lock "refresh_cache-166786ee-4b89-4683-831b-9f3c1c57fb08" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.032259] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.220774] env[62521]: INFO nova.compute.manager [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] [instance: e1ead278-d904-41cc-bdfe-65174e8d6816] Took 1.04 seconds to deallocate network for instance. [ 711.279991] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.553540] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.662026] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.799215] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.829542] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.829796] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.931656] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369715ab-154e-4529-b8f4-6ff82417f4df {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.940036] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96401064-e46d-4b73-8329-c70a813d8857 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.969080] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdebb3a-0964-46b3-8a7d-47a5322f0d24 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.975905] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a07a7d-c6ae-4dba-ab59-934504254f41 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.989546] env[62521]: DEBUG nova.compute.provider_tree [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.165619] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Releasing lock "refresh_cache-166786ee-4b89-4683-831b-9f3c1c57fb08" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.165714] env[62521]: DEBUG nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 712.165893] env[62521]: DEBUG nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.166174] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.181662] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.255018] env[62521]: INFO nova.scheduler.client.report [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Deleted allocations for instance e1ead278-d904-41cc-bdfe-65174e8d6816 [ 712.492741] env[62521]: DEBUG nova.scheduler.client.report [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.684185] env[62521]: DEBUG nova.network.neutron [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.761706] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21a4aee8-800e-480e-87d1-1f455e3aa0b8 tempest-ServersTestBootFromVolume-616516316 tempest-ServersTestBootFromVolume-616516316-project-member] Lock "e1ead278-d904-41cc-bdfe-65174e8d6816" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.170s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.998150] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.972s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.999040] env[62521]: ERROR nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f66095d6-0bf8-4cfc-9921-8fda94e576ea, please check neutron logs for more information. [ 712.999040] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Traceback (most recent call last): [ 712.999040] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.999040] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] self.driver.spawn(context, instance, image_meta, [ 712.999040] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.999040] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.999040] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.999040] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] vm_ref = self.build_virtual_machine(instance, [ 712.999040] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.999040] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.999040] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] for vif in network_info: [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] return self._sync_wrapper(fn, *args, **kwargs) [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] self.wait() [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] self[:] = self._gt.wait() [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] return self._exit_event.wait() [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] result = hub.switch() [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.999448] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] return self.greenlet.switch() [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] result = function(*args, **kwargs) [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] return func(*args, **kwargs) [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] raise e [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] nwinfo = self.network_api.allocate_for_instance( [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] created_port_ids = self._update_ports_for_instance( [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] with excutils.save_and_reraise_exception(): [ 712.999831] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] self.force_reraise() [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] raise self.value [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] updated_port = self._update_port( [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] _ensure_no_port_binding_failure(port) [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] raise exception.PortBindingFailed(port_id=port['id']) [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] nova.exception.PortBindingFailed: Binding failed for port f66095d6-0bf8-4cfc-9921-8fda94e576ea, please check neutron logs for more information. [ 713.000230] env[62521]: ERROR nova.compute.manager [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] [ 713.000551] env[62521]: DEBUG nova.compute.utils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Binding failed for port f66095d6-0bf8-4cfc-9921-8fda94e576ea, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 713.000822] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.567s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.002472] env[62521]: INFO nova.compute.claims [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.005254] env[62521]: DEBUG nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Build of instance cd3d793a-aa56-4f07-a2fa-c56693548dfb was re-scheduled: Binding failed for port f66095d6-0bf8-4cfc-9921-8fda94e576ea, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 713.005713] env[62521]: DEBUG nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 713.005935] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Acquiring lock "refresh_cache-cd3d793a-aa56-4f07-a2fa-c56693548dfb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.006173] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Acquired lock "refresh_cache-cd3d793a-aa56-4f07-a2fa-c56693548dfb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.006441] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.189118] env[62521]: INFO nova.compute.manager [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] [instance: 166786ee-4b89-4683-831b-9f3c1c57fb08] Took 1.02 seconds to deallocate network for instance. [ 713.267280] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.529046] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.632953] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.789746] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.135659] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Releasing lock "refresh_cache-cd3d793a-aa56-4f07-a2fa-c56693548dfb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.135980] env[62521]: DEBUG nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 714.136147] env[62521]: DEBUG nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.136323] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.161138] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.223975] env[62521]: INFO nova.scheduler.client.report [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Deleted allocations for instance 166786ee-4b89-4683-831b-9f3c1c57fb08 [ 714.454423] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbae09e-0cb1-4507-accd-3f92a2f39c6a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.462089] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a363da-a077-4162-be66-08b052a9f7be {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.493240] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c340e43-5a3d-4f45-bc6b-3cdf46a5d3b7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.500412] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952ef86c-0033-4336-946f-489e594da682 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.513769] env[62521]: DEBUG nova.compute.provider_tree [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.665520] env[62521]: DEBUG nova.network.neutron [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.737942] env[62521]: DEBUG oslo_concurrency.lockutils [None req-69fb3c2a-405d-41a8-9f7d-26505eea5e34 tempest-ServerAddressesNegativeTestJSON-1692094859 tempest-ServerAddressesNegativeTestJSON-1692094859-project-member] Lock "166786ee-4b89-4683-831b-9f3c1c57fb08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.537s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.017596] env[62521]: DEBUG nova.scheduler.client.report [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.168397] env[62521]: INFO nova.compute.manager [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] [instance: cd3d793a-aa56-4f07-a2fa-c56693548dfb] Took 1.03 seconds to deallocate network for instance. [ 715.240977] env[62521]: DEBUG nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.523098] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.526623] env[62521]: DEBUG nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.528880] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.106s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.530584] env[62521]: INFO nova.compute.claims [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.763138] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.031169] env[62521]: DEBUG nova.compute.utils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.032669] env[62521]: DEBUG nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 716.033503] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 716.111123] env[62521]: DEBUG nova.policy [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ceb79989e28f4cd0a2cb3671837965f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '531419d03627425da3d3713038b7e54e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 716.201936] env[62521]: INFO nova.scheduler.client.report [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Deleted allocations for instance cd3d793a-aa56-4f07-a2fa-c56693548dfb [ 716.539466] env[62521]: DEBUG nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.612841] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Successfully created port: 967c9d3d-95b2-4703-b878-4787a02da1f5 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.713232] env[62521]: DEBUG oslo_concurrency.lockutils [None req-085462b3-1597-426a-91c8-697050599d2e tempest-FloatingIPsAssociationNegativeTestJSON-1888280080 tempest-FloatingIPsAssociationNegativeTestJSON-1888280080-project-member] Lock "cd3d793a-aa56-4f07-a2fa-c56693548dfb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.198s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.987835] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62150098-68fe-4038-a555-477d573c641d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.995539] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff905614-8c26-46b3-9d4b-d59ba94b5562 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.027979] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826d46c1-28b3-47e0-a81b-704106ec5518 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.034662] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6059d0d5-d954-43b2-8f30-a60e2ba970d0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.052287] env[62521]: DEBUG nova.compute.provider_tree [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.223265] env[62521]: DEBUG nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.561219] env[62521]: DEBUG nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.561637] env[62521]: DEBUG nova.scheduler.client.report [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.602844] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.603157] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.603382] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.603475] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.603639] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.603789] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.604105] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.604971] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.604971] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.604971] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.604971] env[62521]: DEBUG nova.virt.hardware [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.605797] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5528964-00a5-428d-8478-03c3dfd20214 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.614984] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9294d125-ce60-4411-af4b-bf4176175eb0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.701637] env[62521]: DEBUG nova.compute.manager [req-dbd50b07-3d62-429d-8c57-f71c72319b34 req-c05f3228-75bc-4cd7-a9be-87aaedaf92aa service nova] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Received event network-changed-967c9d3d-95b2-4703-b878-4787a02da1f5 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.702227] env[62521]: DEBUG nova.compute.manager [req-dbd50b07-3d62-429d-8c57-f71c72319b34 req-c05f3228-75bc-4cd7-a9be-87aaedaf92aa service nova] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Refreshing instance network info cache due to event network-changed-967c9d3d-95b2-4703-b878-4787a02da1f5. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 717.702468] env[62521]: DEBUG oslo_concurrency.lockutils [req-dbd50b07-3d62-429d-8c57-f71c72319b34 req-c05f3228-75bc-4cd7-a9be-87aaedaf92aa service nova] Acquiring lock "refresh_cache-cf71efc9-cfa1-4856-a733-5434bf4465b1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.702614] env[62521]: DEBUG oslo_concurrency.lockutils [req-dbd50b07-3d62-429d-8c57-f71c72319b34 req-c05f3228-75bc-4cd7-a9be-87aaedaf92aa service nova] Acquired lock "refresh_cache-cf71efc9-cfa1-4856-a733-5434bf4465b1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.702793] env[62521]: DEBUG nova.network.neutron [req-dbd50b07-3d62-429d-8c57-f71c72319b34 req-c05f3228-75bc-4cd7-a9be-87aaedaf92aa service nova] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Refreshing network info cache for port 967c9d3d-95b2-4703-b878-4787a02da1f5 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 717.746467] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.929675] env[62521]: ERROR nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 967c9d3d-95b2-4703-b878-4787a02da1f5, please check neutron logs for more information. [ 717.929675] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 717.929675] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.929675] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 717.929675] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.929675] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 717.929675] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.929675] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 717.929675] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.929675] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 717.929675] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.929675] env[62521]: ERROR nova.compute.manager raise self.value [ 717.929675] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.929675] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 717.929675] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.929675] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 717.930371] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.930371] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 717.930371] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 967c9d3d-95b2-4703-b878-4787a02da1f5, please check neutron logs for more information. [ 717.930371] env[62521]: ERROR nova.compute.manager [ 717.930371] env[62521]: Traceback (most recent call last): [ 717.930371] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 717.930371] env[62521]: listener.cb(fileno) [ 717.930371] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.930371] env[62521]: result = function(*args, **kwargs) [ 717.930371] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.930371] env[62521]: return func(*args, **kwargs) [ 717.930371] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.930371] env[62521]: raise e [ 717.930371] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.930371] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 717.930371] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.930371] env[62521]: created_port_ids = self._update_ports_for_instance( [ 717.930371] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.930371] env[62521]: with excutils.save_and_reraise_exception(): [ 717.930371] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.930371] env[62521]: self.force_reraise() [ 717.930371] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.930371] env[62521]: raise self.value [ 717.930371] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.930371] env[62521]: updated_port = self._update_port( [ 717.930371] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.930371] env[62521]: _ensure_no_port_binding_failure(port) [ 717.930371] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.930371] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 717.931399] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 967c9d3d-95b2-4703-b878-4787a02da1f5, please check neutron logs for more information. [ 717.931399] env[62521]: Removing descriptor: 15 [ 717.931399] env[62521]: ERROR nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 967c9d3d-95b2-4703-b878-4787a02da1f5, please check neutron logs for more information. [ 717.931399] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Traceback (most recent call last): [ 717.931399] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 717.931399] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] yield resources [ 717.931399] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.931399] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] self.driver.spawn(context, instance, image_meta, [ 717.931399] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 717.931399] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.931399] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.931399] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] vm_ref = self.build_virtual_machine(instance, [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] for vif in network_info: [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] return self._sync_wrapper(fn, *args, **kwargs) [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] self.wait() [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] self[:] = self._gt.wait() [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] return self._exit_event.wait() [ 717.931926] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] result = hub.switch() [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] return self.greenlet.switch() [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] result = function(*args, **kwargs) [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] return func(*args, **kwargs) [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] raise e [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] nwinfo = self.network_api.allocate_for_instance( [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.932774] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] created_port_ids = self._update_ports_for_instance( [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] with excutils.save_and_reraise_exception(): [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] self.force_reraise() [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] raise self.value [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] updated_port = self._update_port( [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] _ensure_no_port_binding_failure(port) [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.933853] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] raise exception.PortBindingFailed(port_id=port['id']) [ 717.934421] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] nova.exception.PortBindingFailed: Binding failed for port 967c9d3d-95b2-4703-b878-4787a02da1f5, please check neutron logs for more information. [ 717.934421] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] [ 717.934421] env[62521]: INFO nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Terminating instance [ 717.935850] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquiring lock "refresh_cache-cf71efc9-cfa1-4856-a733-5434bf4465b1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.069589] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.070326] env[62521]: DEBUG nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 718.072695] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.037s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.220815] env[62521]: DEBUG nova.network.neutron [req-dbd50b07-3d62-429d-8c57-f71c72319b34 req-c05f3228-75bc-4cd7-a9be-87aaedaf92aa service nova] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.328490] env[62521]: DEBUG nova.network.neutron [req-dbd50b07-3d62-429d-8c57-f71c72319b34 req-c05f3228-75bc-4cd7-a9be-87aaedaf92aa service nova] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.578011] env[62521]: DEBUG nova.compute.utils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.578984] env[62521]: DEBUG nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 718.579175] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 718.623510] env[62521]: DEBUG nova.policy [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19da1cbfe7094022863f0fa869aeb7dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85061f2b127f4c0190b21f89a80a9294', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 718.830544] env[62521]: DEBUG oslo_concurrency.lockutils [req-dbd50b07-3d62-429d-8c57-f71c72319b34 req-c05f3228-75bc-4cd7-a9be-87aaedaf92aa service nova] Releasing lock "refresh_cache-cf71efc9-cfa1-4856-a733-5434bf4465b1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.830938] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquired lock "refresh_cache-cf71efc9-cfa1-4856-a733-5434bf4465b1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.831145] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.019427] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8f7238-bd30-440e-a82d-3f223e74a61f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.027139] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1b247d-5987-4756-aa6c-6ed779632c51 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.058391] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Successfully created port: 0e835e50-1f4e-4e89-b02e-5ef42824582f {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.060834] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2d9323-372e-40b8-b3d6-ee59de3b90c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.071553] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1759001b-a3d7-4f2f-b633-25fe5728c3f5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.085457] env[62521]: DEBUG nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 719.088053] env[62521]: DEBUG nova.compute.provider_tree [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.356028] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.437583] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.594926] env[62521]: DEBUG nova.scheduler.client.report [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.770227] env[62521]: DEBUG nova.compute.manager [req-5a70534a-3e36-49dd-8b37-a0c519e59139 req-356ad68f-a796-4c83-a8f9-3c51d98579b1 service nova] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Received event network-vif-deleted-967c9d3d-95b2-4703-b878-4787a02da1f5 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.945772] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Releasing lock "refresh_cache-cf71efc9-cfa1-4856-a733-5434bf4465b1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.946157] env[62521]: DEBUG nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 719.946355] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 719.947263] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18c36b8c-5c59-4da4-afeb-606c35cb0094 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.955852] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5db2f8e-62e7-452f-9586-38132ed15f72 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.977894] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cf71efc9-cfa1-4856-a733-5434bf4465b1 could not be found. [ 719.978053] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 719.978246] env[62521]: INFO nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 719.978491] env[62521]: DEBUG oslo.service.loopingcall [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.978750] env[62521]: DEBUG nova.compute.manager [-] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.978800] env[62521]: DEBUG nova.network.neutron [-] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.998961] env[62521]: DEBUG nova.network.neutron [-] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.102828] env[62521]: DEBUG nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 720.104352] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.032s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.105125] env[62521]: ERROR nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 67d58d1f-300c-497f-af25-825bc30fae23, please check neutron logs for more information. [ 720.105125] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] Traceback (most recent call last): [ 720.105125] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.105125] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] self.driver.spawn(context, instance, image_meta, [ 720.105125] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 720.105125] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.105125] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.105125] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] vm_ref = self.build_virtual_machine(instance, [ 720.105125] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.105125] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.105125] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] for vif in network_info: [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] return self._sync_wrapper(fn, *args, **kwargs) [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] self.wait() [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] self[:] = self._gt.wait() [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] return self._exit_event.wait() [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] result = hub.switch() [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 720.105508] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] return self.greenlet.switch() [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] result = function(*args, **kwargs) [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] return func(*args, **kwargs) [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] raise e [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] nwinfo = self.network_api.allocate_for_instance( [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] created_port_ids = self._update_ports_for_instance( [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] with excutils.save_and_reraise_exception(): [ 720.105962] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] self.force_reraise() [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] raise self.value [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] updated_port = self._update_port( [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] _ensure_no_port_binding_failure(port) [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] raise exception.PortBindingFailed(port_id=port['id']) [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] nova.exception.PortBindingFailed: Binding failed for port 67d58d1f-300c-497f-af25-825bc30fae23, please check neutron logs for more information. [ 720.106337] env[62521]: ERROR nova.compute.manager [instance: a4965864-0127-4842-9478-76e369e2ecbe] [ 720.106638] env[62521]: DEBUG nova.compute.utils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Binding failed for port 67d58d1f-300c-497f-af25-825bc30fae23, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 720.107744] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.306s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.109160] env[62521]: INFO nova.compute.claims [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.113478] env[62521]: DEBUG nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Build of instance a4965864-0127-4842-9478-76e369e2ecbe was re-scheduled: Binding failed for port 67d58d1f-300c-497f-af25-825bc30fae23, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 720.114436] env[62521]: DEBUG nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 720.115053] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Acquiring lock "refresh_cache-a4965864-0127-4842-9478-76e369e2ecbe" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.115223] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Acquired lock "refresh_cache-a4965864-0127-4842-9478-76e369e2ecbe" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.115385] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.136457] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.136719] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.136881] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.137366] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.137567] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.137725] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.137937] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.138291] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.139022] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.139022] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.139022] env[62521]: DEBUG nova.virt.hardware [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.140310] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a8b0aa-8e51-474a-a8d8-de4c94ec90f1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.150246] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d973fba0-6963-45d6-b66a-883b61ff275a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.463270] env[62521]: ERROR nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0e835e50-1f4e-4e89-b02e-5ef42824582f, please check neutron logs for more information. [ 720.463270] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 720.463270] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.463270] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 720.463270] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.463270] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 720.463270] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.463270] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 720.463270] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.463270] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 720.463270] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.463270] env[62521]: ERROR nova.compute.manager raise self.value [ 720.463270] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.463270] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 720.463270] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.463270] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 720.463954] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.463954] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 720.463954] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0e835e50-1f4e-4e89-b02e-5ef42824582f, please check neutron logs for more information. [ 720.463954] env[62521]: ERROR nova.compute.manager [ 720.463954] env[62521]: Traceback (most recent call last): [ 720.463954] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 720.463954] env[62521]: listener.cb(fileno) [ 720.463954] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.463954] env[62521]: result = function(*args, **kwargs) [ 720.463954] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 720.463954] env[62521]: return func(*args, **kwargs) [ 720.463954] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.463954] env[62521]: raise e [ 720.463954] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.463954] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 720.463954] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.463954] env[62521]: created_port_ids = self._update_ports_for_instance( [ 720.463954] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.463954] env[62521]: with excutils.save_and_reraise_exception(): [ 720.463954] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.463954] env[62521]: self.force_reraise() [ 720.463954] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.463954] env[62521]: raise self.value [ 720.463954] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.463954] env[62521]: updated_port = self._update_port( [ 720.463954] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.463954] env[62521]: _ensure_no_port_binding_failure(port) [ 720.463954] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.463954] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 720.464673] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 0e835e50-1f4e-4e89-b02e-5ef42824582f, please check neutron logs for more information. [ 720.464673] env[62521]: Removing descriptor: 15 [ 720.464673] env[62521]: ERROR nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0e835e50-1f4e-4e89-b02e-5ef42824582f, please check neutron logs for more information. [ 720.464673] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Traceback (most recent call last): [ 720.464673] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 720.464673] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] yield resources [ 720.464673] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.464673] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] self.driver.spawn(context, instance, image_meta, [ 720.464673] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 720.464673] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.464673] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.464673] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] vm_ref = self.build_virtual_machine(instance, [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] for vif in network_info: [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] return self._sync_wrapper(fn, *args, **kwargs) [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] self.wait() [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] self[:] = self._gt.wait() [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] return self._exit_event.wait() [ 720.465009] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] result = hub.switch() [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] return self.greenlet.switch() [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] result = function(*args, **kwargs) [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] return func(*args, **kwargs) [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] raise e [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] nwinfo = self.network_api.allocate_for_instance( [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.465332] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] created_port_ids = self._update_ports_for_instance( [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] with excutils.save_and_reraise_exception(): [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] self.force_reraise() [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] raise self.value [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] updated_port = self._update_port( [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] _ensure_no_port_binding_failure(port) [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.465661] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] raise exception.PortBindingFailed(port_id=port['id']) [ 720.466104] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] nova.exception.PortBindingFailed: Binding failed for port 0e835e50-1f4e-4e89-b02e-5ef42824582f, please check neutron logs for more information. [ 720.466104] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] [ 720.466104] env[62521]: INFO nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Terminating instance [ 720.468489] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Acquiring lock "refresh_cache-d7575b46-ecf9-4154-92b1-86119e727dda" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.468660] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Acquired lock "refresh_cache-d7575b46-ecf9-4154-92b1-86119e727dda" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.468825] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.501630] env[62521]: DEBUG nova.network.neutron [-] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.631503] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.718744] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.993353] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.003787] env[62521]: INFO nova.compute.manager [-] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Took 1.02 seconds to deallocate network for instance. [ 721.006825] env[62521]: DEBUG nova.compute.claims [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.007061] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.107685] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.224047] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Releasing lock "refresh_cache-a4965864-0127-4842-9478-76e369e2ecbe" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.224047] env[62521]: DEBUG nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 721.224047] env[62521]: DEBUG nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 721.224047] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.242049] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.482491] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651be176-6ae0-4cf2-b8a2-192894788d82 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.490666] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f4c27a-433f-409b-b3e5-4345273f79c0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.520298] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd2516b-898c-4698-aba9-070588ea9a98 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.527845] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a0864d-7234-49c2-b56b-38dec5ed7ff5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.541155] env[62521]: DEBUG nova.compute.provider_tree [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.611450] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Releasing lock "refresh_cache-d7575b46-ecf9-4154-92b1-86119e727dda" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.611871] env[62521]: DEBUG nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 721.612127] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 721.612440] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e58b1f02-be4c-4acc-b5d7-2f739e34ffe1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.621972] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47567381-fa4d-4763-89d2-6b5444d087b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.642926] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7575b46-ecf9-4154-92b1-86119e727dda could not be found. [ 721.643164] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 721.643348] env[62521]: INFO nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Took 0.03 seconds to destroy the instance on the hypervisor. [ 721.643613] env[62521]: DEBUG oslo.service.loopingcall [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.643790] env[62521]: DEBUG nova.compute.manager [-] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 721.643881] env[62521]: DEBUG nova.network.neutron [-] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.663103] env[62521]: DEBUG nova.network.neutron [-] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.750158] env[62521]: DEBUG nova.network.neutron [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.931200] env[62521]: DEBUG nova.compute.manager [req-d45190bb-0a4b-4fad-b1f9-dcbcac62024a req-77f68ba3-bdd2-4a35-bf9c-50c6064a1923 service nova] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Received event network-changed-0e835e50-1f4e-4e89-b02e-5ef42824582f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.931200] env[62521]: DEBUG nova.compute.manager [req-d45190bb-0a4b-4fad-b1f9-dcbcac62024a req-77f68ba3-bdd2-4a35-bf9c-50c6064a1923 service nova] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Refreshing instance network info cache due to event network-changed-0e835e50-1f4e-4e89-b02e-5ef42824582f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 721.931200] env[62521]: DEBUG oslo_concurrency.lockutils [req-d45190bb-0a4b-4fad-b1f9-dcbcac62024a req-77f68ba3-bdd2-4a35-bf9c-50c6064a1923 service nova] Acquiring lock "refresh_cache-d7575b46-ecf9-4154-92b1-86119e727dda" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.931200] env[62521]: DEBUG oslo_concurrency.lockutils [req-d45190bb-0a4b-4fad-b1f9-dcbcac62024a req-77f68ba3-bdd2-4a35-bf9c-50c6064a1923 service nova] Acquired lock "refresh_cache-d7575b46-ecf9-4154-92b1-86119e727dda" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.931200] env[62521]: DEBUG nova.network.neutron [req-d45190bb-0a4b-4fad-b1f9-dcbcac62024a req-77f68ba3-bdd2-4a35-bf9c-50c6064a1923 service nova] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Refreshing network info cache for port 0e835e50-1f4e-4e89-b02e-5ef42824582f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.046152] env[62521]: DEBUG nova.scheduler.client.report [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.166254] env[62521]: DEBUG nova.network.neutron [-] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.253153] env[62521]: INFO nova.compute.manager [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] [instance: a4965864-0127-4842-9478-76e369e2ecbe] Took 1.03 seconds to deallocate network for instance. [ 722.465934] env[62521]: DEBUG nova.network.neutron [req-d45190bb-0a4b-4fad-b1f9-dcbcac62024a req-77f68ba3-bdd2-4a35-bf9c-50c6064a1923 service nova] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.542257] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquiring lock "4afe09db-6c01-444f-a127-6e1f97794544" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.542577] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lock "4afe09db-6c01-444f-a127-6e1f97794544" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.552529] env[62521]: DEBUG nova.network.neutron [req-d45190bb-0a4b-4fad-b1f9-dcbcac62024a req-77f68ba3-bdd2-4a35-bf9c-50c6064a1923 service nova] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.552847] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.553326] env[62521]: DEBUG nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.556721] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.714s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.558199] env[62521]: INFO nova.compute.claims [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 722.669673] env[62521]: INFO nova.compute.manager [-] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Took 1.03 seconds to deallocate network for instance. [ 722.672082] env[62521]: DEBUG nova.compute.claims [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 722.672229] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.058638] env[62521]: DEBUG nova.compute.utils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.060592] env[62521]: DEBUG oslo_concurrency.lockutils [req-d45190bb-0a4b-4fad-b1f9-dcbcac62024a req-77f68ba3-bdd2-4a35-bf9c-50c6064a1923 service nova] Releasing lock "refresh_cache-d7575b46-ecf9-4154-92b1-86119e727dda" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.060888] env[62521]: DEBUG nova.compute.manager [req-d45190bb-0a4b-4fad-b1f9-dcbcac62024a req-77f68ba3-bdd2-4a35-bf9c-50c6064a1923 service nova] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Received event network-vif-deleted-0e835e50-1f4e-4e89-b02e-5ef42824582f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.061506] env[62521]: DEBUG nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 723.061726] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 723.130113] env[62521]: DEBUG nova.policy [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '39a60744d425428c88a71cc29061a1cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '941c76a8bff8493288a960a452bb1343', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.307234] env[62521]: INFO nova.scheduler.client.report [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Deleted allocations for instance a4965864-0127-4842-9478-76e369e2ecbe [ 723.497939] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Successfully created port: c621baa7-5581-4e73-a0df-0b154224d696 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.562735] env[62521]: DEBUG nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.820795] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c67fc6fd-262d-450b-9617-5c8e87a69ba5 tempest-ServersV294TestFqdnHostnames-620975717 tempest-ServersV294TestFqdnHostnames-620975717-project-member] Lock "a4965864-0127-4842-9478-76e369e2ecbe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.943s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.914858] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ea391e-9fe8-4d9d-baae-5048f07d2ef1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.924613] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcdc059-f853-4181-9460-f953cb012a3b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.955063] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f099585-970a-4ba7-8d1a-8d934d5e8d3f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.962599] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64106081-e6e4-4cbc-9903-ed6f8eb8683c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.976439] env[62521]: DEBUG nova.compute.provider_tree [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.328262] env[62521]: DEBUG nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 724.481012] env[62521]: DEBUG nova.compute.manager [req-e6938bcf-25be-4d28-af59-392f20b758ae req-be8a9281-5479-404c-8b7c-0a13e0be8319 service nova] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Received event network-changed-c621baa7-5581-4e73-a0df-0b154224d696 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.481262] env[62521]: DEBUG nova.compute.manager [req-e6938bcf-25be-4d28-af59-392f20b758ae req-be8a9281-5479-404c-8b7c-0a13e0be8319 service nova] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Refreshing instance network info cache due to event network-changed-c621baa7-5581-4e73-a0df-0b154224d696. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 724.481560] env[62521]: DEBUG oslo_concurrency.lockutils [req-e6938bcf-25be-4d28-af59-392f20b758ae req-be8a9281-5479-404c-8b7c-0a13e0be8319 service nova] Acquiring lock "refresh_cache-0d72d988-f96b-428c-89bf-a96572f71db5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.481718] env[62521]: DEBUG oslo_concurrency.lockutils [req-e6938bcf-25be-4d28-af59-392f20b758ae req-be8a9281-5479-404c-8b7c-0a13e0be8319 service nova] Acquired lock "refresh_cache-0d72d988-f96b-428c-89bf-a96572f71db5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.481889] env[62521]: DEBUG nova.network.neutron [req-e6938bcf-25be-4d28-af59-392f20b758ae req-be8a9281-5479-404c-8b7c-0a13e0be8319 service nova] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Refreshing network info cache for port c621baa7-5581-4e73-a0df-0b154224d696 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 724.483584] env[62521]: DEBUG nova.scheduler.client.report [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.578019] env[62521]: DEBUG nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.605494] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.605494] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.605638] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.606651] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.606651] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.606651] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.606854] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.607040] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.607295] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.607494] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.607701] env[62521]: DEBUG nova.virt.hardware [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.608806] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9ebff8-43ef-437e-84ed-0562a5ffa97e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.618183] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfb5741-ed02-4443-957f-16b5ae0c5a7f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.657071] env[62521]: ERROR nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c621baa7-5581-4e73-a0df-0b154224d696, please check neutron logs for more information. [ 724.657071] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 724.657071] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.657071] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 724.657071] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.657071] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 724.657071] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.657071] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 724.657071] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.657071] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 724.657071] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.657071] env[62521]: ERROR nova.compute.manager raise self.value [ 724.657071] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.657071] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 724.657071] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.657071] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 724.657592] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.657592] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 724.657592] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c621baa7-5581-4e73-a0df-0b154224d696, please check neutron logs for more information. [ 724.657592] env[62521]: ERROR nova.compute.manager [ 724.657592] env[62521]: Traceback (most recent call last): [ 724.657592] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 724.657592] env[62521]: listener.cb(fileno) [ 724.657592] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.657592] env[62521]: result = function(*args, **kwargs) [ 724.657592] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.657592] env[62521]: return func(*args, **kwargs) [ 724.657592] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.657592] env[62521]: raise e [ 724.657592] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.657592] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 724.657592] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.657592] env[62521]: created_port_ids = self._update_ports_for_instance( [ 724.657592] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.657592] env[62521]: with excutils.save_and_reraise_exception(): [ 724.657592] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.657592] env[62521]: self.force_reraise() [ 724.657592] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.657592] env[62521]: raise self.value [ 724.657592] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.657592] env[62521]: updated_port = self._update_port( [ 724.657592] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.657592] env[62521]: _ensure_no_port_binding_failure(port) [ 724.657592] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.657592] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 724.658532] env[62521]: nova.exception.PortBindingFailed: Binding failed for port c621baa7-5581-4e73-a0df-0b154224d696, please check neutron logs for more information. [ 724.658532] env[62521]: Removing descriptor: 15 [ 724.658532] env[62521]: ERROR nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c621baa7-5581-4e73-a0df-0b154224d696, please check neutron logs for more information. [ 724.658532] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Traceback (most recent call last): [ 724.658532] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 724.658532] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] yield resources [ 724.658532] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.658532] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] self.driver.spawn(context, instance, image_meta, [ 724.658532] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 724.658532] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.658532] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.658532] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] vm_ref = self.build_virtual_machine(instance, [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] for vif in network_info: [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] return self._sync_wrapper(fn, *args, **kwargs) [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] self.wait() [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] self[:] = self._gt.wait() [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] return self._exit_event.wait() [ 724.658897] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] result = hub.switch() [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] return self.greenlet.switch() [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] result = function(*args, **kwargs) [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] return func(*args, **kwargs) [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] raise e [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] nwinfo = self.network_api.allocate_for_instance( [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.659300] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] created_port_ids = self._update_ports_for_instance( [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] with excutils.save_and_reraise_exception(): [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] self.force_reraise() [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] raise self.value [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] updated_port = self._update_port( [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] _ensure_no_port_binding_failure(port) [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.659694] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] raise exception.PortBindingFailed(port_id=port['id']) [ 724.660082] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] nova.exception.PortBindingFailed: Binding failed for port c621baa7-5581-4e73-a0df-0b154224d696, please check neutron logs for more information. [ 724.660082] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] [ 724.660082] env[62521]: INFO nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Terminating instance [ 724.661462] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "refresh_cache-0d72d988-f96b-428c-89bf-a96572f71db5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.850603] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.990022] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.990322] env[62521]: DEBUG nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 724.993162] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.198s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.994668] env[62521]: INFO nova.compute.claims [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.013569] env[62521]: DEBUG nova.network.neutron [req-e6938bcf-25be-4d28-af59-392f20b758ae req-be8a9281-5479-404c-8b7c-0a13e0be8319 service nova] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.113458] env[62521]: DEBUG nova.network.neutron [req-e6938bcf-25be-4d28-af59-392f20b758ae req-be8a9281-5479-404c-8b7c-0a13e0be8319 service nova] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.502056] env[62521]: DEBUG nova.compute.utils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.503568] env[62521]: DEBUG nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 725.507156] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 725.557525] env[62521]: DEBUG nova.policy [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '39a60744d425428c88a71cc29061a1cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '941c76a8bff8493288a960a452bb1343', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 725.618288] env[62521]: DEBUG oslo_concurrency.lockutils [req-e6938bcf-25be-4d28-af59-392f20b758ae req-be8a9281-5479-404c-8b7c-0a13e0be8319 service nova] Releasing lock "refresh_cache-0d72d988-f96b-428c-89bf-a96572f71db5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.619445] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquired lock "refresh_cache-0d72d988-f96b-428c-89bf-a96572f71db5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.620260] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.010969] env[62521]: DEBUG nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 726.067478] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Successfully created port: 6a4810e6-a03d-46be-bc14-9448caf0aebb {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 726.143431] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.242247] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.421383] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337cddb9-9d08-461e-b48b-6da2810d45c1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.432019] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e21f08-40be-493c-9a08-8f58af5eb23f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.460351] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0238a600-6683-4860-a6ea-696b88c6d0a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.467529] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e6523a-36aa-4c32-a682-14487b614821 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.481614] env[62521]: DEBUG nova.compute.provider_tree [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.519027] env[62521]: DEBUG nova.compute.manager [req-6034ed52-d8a3-445b-a9b3-f9b7a1a33b08 req-1b40ad56-49c3-434b-9f87-b2af1f989a25 service nova] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Received event network-vif-deleted-c621baa7-5581-4e73-a0df-0b154224d696 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.747932] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Releasing lock "refresh_cache-0d72d988-f96b-428c-89bf-a96572f71db5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.747932] env[62521]: DEBUG nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.747932] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 726.747932] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cdf5f9ce-ac15-4c55-a156-f9d930528ff3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.757033] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e927bb-97e5-430a-b64a-dfd3ce4b39f2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.776277] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0d72d988-f96b-428c-89bf-a96572f71db5 could not be found. [ 726.776450] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 726.776731] env[62521]: INFO nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 726.777043] env[62521]: DEBUG oslo.service.loopingcall [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.777309] env[62521]: DEBUG nova.compute.manager [-] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.777446] env[62521]: DEBUG nova.network.neutron [-] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.793139] env[62521]: DEBUG nova.network.neutron [-] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.985054] env[62521]: DEBUG nova.scheduler.client.report [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.019461] env[62521]: DEBUG nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 727.044628] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 727.044875] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 727.045118] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 727.045322] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 727.045470] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 727.045617] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 727.045841] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 727.046013] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 727.046218] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 727.046399] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 727.046551] env[62521]: DEBUG nova.virt.hardware [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 727.047413] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9317482-4d4e-46f9-aea5-2233432589c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.055326] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b7d9f3-fa09-41fb-81d5-98104ad8af96 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.126034] env[62521]: ERROR nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6a4810e6-a03d-46be-bc14-9448caf0aebb, please check neutron logs for more information. [ 727.126034] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 727.126034] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.126034] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 727.126034] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.126034] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 727.126034] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.126034] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 727.126034] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.126034] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 727.126034] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.126034] env[62521]: ERROR nova.compute.manager raise self.value [ 727.126034] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.126034] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 727.126034] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.126034] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 727.126563] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.126563] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 727.126563] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6a4810e6-a03d-46be-bc14-9448caf0aebb, please check neutron logs for more information. [ 727.126563] env[62521]: ERROR nova.compute.manager [ 727.126563] env[62521]: Traceback (most recent call last): [ 727.126563] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 727.126563] env[62521]: listener.cb(fileno) [ 727.126563] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.126563] env[62521]: result = function(*args, **kwargs) [ 727.126563] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 727.126563] env[62521]: return func(*args, **kwargs) [ 727.126563] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.126563] env[62521]: raise e [ 727.126563] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.126563] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 727.126563] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.126563] env[62521]: created_port_ids = self._update_ports_for_instance( [ 727.126563] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.126563] env[62521]: with excutils.save_and_reraise_exception(): [ 727.126563] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.126563] env[62521]: self.force_reraise() [ 727.126563] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.126563] env[62521]: raise self.value [ 727.126563] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.126563] env[62521]: updated_port = self._update_port( [ 727.126563] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.126563] env[62521]: _ensure_no_port_binding_failure(port) [ 727.126563] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.126563] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 727.127592] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 6a4810e6-a03d-46be-bc14-9448caf0aebb, please check neutron logs for more information. [ 727.127592] env[62521]: Removing descriptor: 15 [ 727.127592] env[62521]: ERROR nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6a4810e6-a03d-46be-bc14-9448caf0aebb, please check neutron logs for more information. [ 727.127592] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Traceback (most recent call last): [ 727.127592] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 727.127592] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] yield resources [ 727.127592] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 727.127592] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] self.driver.spawn(context, instance, image_meta, [ 727.127592] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 727.127592] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] self._vmops.spawn(context, instance, image_meta, injected_files, [ 727.127592] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 727.127592] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] vm_ref = self.build_virtual_machine(instance, [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] vif_infos = vmwarevif.get_vif_info(self._session, [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] for vif in network_info: [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] return self._sync_wrapper(fn, *args, **kwargs) [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] self.wait() [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] self[:] = self._gt.wait() [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] return self._exit_event.wait() [ 727.127982] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] result = hub.switch() [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] return self.greenlet.switch() [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] result = function(*args, **kwargs) [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] return func(*args, **kwargs) [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] raise e [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] nwinfo = self.network_api.allocate_for_instance( [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.128404] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] created_port_ids = self._update_ports_for_instance( [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] with excutils.save_and_reraise_exception(): [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] self.force_reraise() [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] raise self.value [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] updated_port = self._update_port( [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] _ensure_no_port_binding_failure(port) [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.128782] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] raise exception.PortBindingFailed(port_id=port['id']) [ 727.129156] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] nova.exception.PortBindingFailed: Binding failed for port 6a4810e6-a03d-46be-bc14-9448caf0aebb, please check neutron logs for more information. [ 727.129156] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] [ 727.129156] env[62521]: INFO nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Terminating instance [ 727.129156] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "refresh_cache-884d2f4f-c0a2-4d08-962e-d6347dc9f678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.129156] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquired lock "refresh_cache-884d2f4f-c0a2-4d08-962e-d6347dc9f678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.129156] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.295897] env[62521]: DEBUG nova.network.neutron [-] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.491515] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.492051] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.494553] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.695s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.497671] env[62521]: INFO nova.compute.claims [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.655621] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.736574] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.799099] env[62521]: INFO nova.compute.manager [-] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Took 1.02 seconds to deallocate network for instance. [ 727.801277] env[62521]: DEBUG nova.compute.claims [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 727.801451] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.002049] env[62521]: DEBUG nova.compute.utils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.005343] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 728.005520] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 728.051742] env[62521]: DEBUG nova.policy [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8c0ef26e3b5414aa6f7ad999051f076', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2a99691b0c9a407fb90fd95bf9232c48', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 728.239099] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Releasing lock "refresh_cache-884d2f4f-c0a2-4d08-962e-d6347dc9f678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.239555] env[62521]: DEBUG nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.239750] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.240186] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ebf34e1f-97bb-4c59-9152-ab4e4246af20 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.249674] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41893f7-5dd9-4f79-8665-47537b1b095c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.271854] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 884d2f4f-c0a2-4d08-962e-d6347dc9f678 could not be found. [ 728.272091] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.272283] env[62521]: INFO nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Took 0.03 seconds to destroy the instance on the hypervisor. [ 728.272521] env[62521]: DEBUG oslo.service.loopingcall [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.272743] env[62521]: DEBUG nova.compute.manager [-] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.272835] env[62521]: DEBUG nova.network.neutron [-] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.287336] env[62521]: DEBUG nova.network.neutron [-] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.345274] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Successfully created port: 1f89168c-d105-46ac-9320-34456e056199 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 728.506965] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 728.578269] env[62521]: DEBUG nova.compute.manager [req-5fb63755-34f7-4e61-a5fd-306fb6f05e87 req-e45e258e-f3da-4b25-8c13-33c8b4580f94 service nova] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Received event network-changed-6a4810e6-a03d-46be-bc14-9448caf0aebb {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.578466] env[62521]: DEBUG nova.compute.manager [req-5fb63755-34f7-4e61-a5fd-306fb6f05e87 req-e45e258e-f3da-4b25-8c13-33c8b4580f94 service nova] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Refreshing instance network info cache due to event network-changed-6a4810e6-a03d-46be-bc14-9448caf0aebb. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 728.578682] env[62521]: DEBUG oslo_concurrency.lockutils [req-5fb63755-34f7-4e61-a5fd-306fb6f05e87 req-e45e258e-f3da-4b25-8c13-33c8b4580f94 service nova] Acquiring lock "refresh_cache-884d2f4f-c0a2-4d08-962e-d6347dc9f678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.578821] env[62521]: DEBUG oslo_concurrency.lockutils [req-5fb63755-34f7-4e61-a5fd-306fb6f05e87 req-e45e258e-f3da-4b25-8c13-33c8b4580f94 service nova] Acquired lock "refresh_cache-884d2f4f-c0a2-4d08-962e-d6347dc9f678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.578980] env[62521]: DEBUG nova.network.neutron [req-5fb63755-34f7-4e61-a5fd-306fb6f05e87 req-e45e258e-f3da-4b25-8c13-33c8b4580f94 service nova] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Refreshing network info cache for port 6a4810e6-a03d-46be-bc14-9448caf0aebb {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 728.789470] env[62521]: DEBUG nova.network.neutron [-] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.886797] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95b4ec9-48f4-4c0c-bff6-ce48c41e2fd0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.894320] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b5eb93-2141-481c-a897-6bf956320811 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.924438] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00da3205-bf83-4251-b514-f21ea04dfb2c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.931195] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673a9528-27df-49ec-aacb-330d668c6e9c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.944595] env[62521]: DEBUG nova.compute.provider_tree [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.097056] env[62521]: DEBUG nova.network.neutron [req-5fb63755-34f7-4e61-a5fd-306fb6f05e87 req-e45e258e-f3da-4b25-8c13-33c8b4580f94 service nova] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.179226] env[62521]: DEBUG nova.network.neutron [req-5fb63755-34f7-4e61-a5fd-306fb6f05e87 req-e45e258e-f3da-4b25-8c13-33c8b4580f94 service nova] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.294009] env[62521]: INFO nova.compute.manager [-] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Took 1.02 seconds to deallocate network for instance. [ 729.296327] env[62521]: DEBUG nova.compute.claims [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 729.296499] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.443290] env[62521]: ERROR nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f89168c-d105-46ac-9320-34456e056199, please check neutron logs for more information. [ 729.443290] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 729.443290] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.443290] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 729.443290] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.443290] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 729.443290] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.443290] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 729.443290] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.443290] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 729.443290] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.443290] env[62521]: ERROR nova.compute.manager raise self.value [ 729.443290] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.443290] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 729.443290] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.443290] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 729.443832] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.443832] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 729.443832] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f89168c-d105-46ac-9320-34456e056199, please check neutron logs for more information. [ 729.443832] env[62521]: ERROR nova.compute.manager [ 729.443832] env[62521]: Traceback (most recent call last): [ 729.443832] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 729.443832] env[62521]: listener.cb(fileno) [ 729.443832] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.443832] env[62521]: result = function(*args, **kwargs) [ 729.443832] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.443832] env[62521]: return func(*args, **kwargs) [ 729.443832] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.443832] env[62521]: raise e [ 729.443832] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.443832] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 729.443832] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.443832] env[62521]: created_port_ids = self._update_ports_for_instance( [ 729.443832] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.443832] env[62521]: with excutils.save_and_reraise_exception(): [ 729.443832] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.443832] env[62521]: self.force_reraise() [ 729.443832] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.443832] env[62521]: raise self.value [ 729.443832] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.443832] env[62521]: updated_port = self._update_port( [ 729.443832] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.443832] env[62521]: _ensure_no_port_binding_failure(port) [ 729.443832] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.443832] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 729.444705] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 1f89168c-d105-46ac-9320-34456e056199, please check neutron logs for more information. [ 729.444705] env[62521]: Removing descriptor: 15 [ 729.447494] env[62521]: DEBUG nova.scheduler.client.report [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.526203] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.552215] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.552598] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.552767] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.552956] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.553113] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.553262] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.553545] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.553711] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.553876] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.554044] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.554219] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.555185] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9825f0f-6281-49eb-bbb4-53c9fc96290d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.563139] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7d8076-2d29-4321-af9b-a90e2001e415 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.576683] env[62521]: ERROR nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f89168c-d105-46ac-9320-34456e056199, please check neutron logs for more information. [ 729.576683] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Traceback (most recent call last): [ 729.576683] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 729.576683] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] yield resources [ 729.576683] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.576683] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] self.driver.spawn(context, instance, image_meta, [ 729.576683] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 729.576683] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.576683] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.576683] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] vm_ref = self.build_virtual_machine(instance, [ 729.576683] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] for vif in network_info: [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] return self._sync_wrapper(fn, *args, **kwargs) [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] self.wait() [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] self[:] = self._gt.wait() [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] return self._exit_event.wait() [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 729.577064] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] current.throw(*self._exc) [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] result = function(*args, **kwargs) [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] return func(*args, **kwargs) [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] raise e [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] nwinfo = self.network_api.allocate_for_instance( [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] created_port_ids = self._update_ports_for_instance( [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] with excutils.save_and_reraise_exception(): [ 729.577444] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] self.force_reraise() [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] raise self.value [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] updated_port = self._update_port( [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] _ensure_no_port_binding_failure(port) [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] raise exception.PortBindingFailed(port_id=port['id']) [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] nova.exception.PortBindingFailed: Binding failed for port 1f89168c-d105-46ac-9320-34456e056199, please check neutron logs for more information. [ 729.577831] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] [ 729.577831] env[62521]: INFO nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Terminating instance [ 729.578992] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "refresh_cache-5fa1c73e-3074-4fe0-8b4d-926bf02f1453" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.579171] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquired lock "refresh_cache-5fa1c73e-3074-4fe0-8b4d-926bf02f1453" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.579338] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 729.680626] env[62521]: DEBUG oslo_concurrency.lockutils [req-5fb63755-34f7-4e61-a5fd-306fb6f05e87 req-e45e258e-f3da-4b25-8c13-33c8b4580f94 service nova] Releasing lock "refresh_cache-884d2f4f-c0a2-4d08-962e-d6347dc9f678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.680907] env[62521]: DEBUG nova.compute.manager [req-5fb63755-34f7-4e61-a5fd-306fb6f05e87 req-e45e258e-f3da-4b25-8c13-33c8b4580f94 service nova] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Received event network-vif-deleted-6a4810e6-a03d-46be-bc14-9448caf0aebb {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.953791] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.954312] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.957039] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.167s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.958876] env[62521]: INFO nova.compute.claims [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.095966] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.189278] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.465863] env[62521]: DEBUG nova.compute.utils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.470063] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 730.470063] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 730.527060] env[62521]: DEBUG nova.policy [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8c0ef26e3b5414aa6f7ad999051f076', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2a99691b0c9a407fb90fd95bf9232c48', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 730.612188] env[62521]: DEBUG nova.compute.manager [req-85f3a95e-37eb-41db-a5ae-491ec60a518c req-e2ba22f1-11aa-435f-a3d5-3fb884eb93c5 service nova] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Received event network-changed-1f89168c-d105-46ac-9320-34456e056199 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.612401] env[62521]: DEBUG nova.compute.manager [req-85f3a95e-37eb-41db-a5ae-491ec60a518c req-e2ba22f1-11aa-435f-a3d5-3fb884eb93c5 service nova] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Refreshing instance network info cache due to event network-changed-1f89168c-d105-46ac-9320-34456e056199. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 730.612587] env[62521]: DEBUG oslo_concurrency.lockutils [req-85f3a95e-37eb-41db-a5ae-491ec60a518c req-e2ba22f1-11aa-435f-a3d5-3fb884eb93c5 service nova] Acquiring lock "refresh_cache-5fa1c73e-3074-4fe0-8b4d-926bf02f1453" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.691499] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Releasing lock "refresh_cache-5fa1c73e-3074-4fe0-8b4d-926bf02f1453" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.691925] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 730.692261] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 730.692578] env[62521]: DEBUG oslo_concurrency.lockutils [req-85f3a95e-37eb-41db-a5ae-491ec60a518c req-e2ba22f1-11aa-435f-a3d5-3fb884eb93c5 service nova] Acquired lock "refresh_cache-5fa1c73e-3074-4fe0-8b4d-926bf02f1453" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.692749] env[62521]: DEBUG nova.network.neutron [req-85f3a95e-37eb-41db-a5ae-491ec60a518c req-e2ba22f1-11aa-435f-a3d5-3fb884eb93c5 service nova] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Refreshing network info cache for port 1f89168c-d105-46ac-9320-34456e056199 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 730.694393] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a12285e4-8d04-4588-abd1-9815fc7b1059 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.703685] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4616b59-d037-4376-a654-3507c9d23409 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.724736] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5fa1c73e-3074-4fe0-8b4d-926bf02f1453 could not be found. [ 730.724955] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 730.725162] env[62521]: INFO nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Took 0.03 seconds to destroy the instance on the hypervisor. [ 730.725398] env[62521]: DEBUG oslo.service.loopingcall [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.725609] env[62521]: DEBUG nova.compute.manager [-] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.725696] env[62521]: DEBUG nova.network.neutron [-] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.743946] env[62521]: DEBUG nova.network.neutron [-] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.900218] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Successfully created port: 444157f5-a8b1-4bc7-891f-25e923aaf7c2 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.972610] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 731.217721] env[62521]: DEBUG nova.network.neutron [req-85f3a95e-37eb-41db-a5ae-491ec60a518c req-e2ba22f1-11aa-435f-a3d5-3fb884eb93c5 service nova] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.246970] env[62521]: DEBUG nova.network.neutron [-] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.298268] env[62521]: DEBUG nova.network.neutron [req-85f3a95e-37eb-41db-a5ae-491ec60a518c req-e2ba22f1-11aa-435f-a3d5-3fb884eb93c5 service nova] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.347017] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a89b23-e5e2-4ed0-b32e-ccf17e8fc260 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.355148] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777a3ea0-a250-499e-9cb1-3480d4abe140 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.386484] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446f499f-63cb-4b73-9bfb-5cf190bc48fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.393645] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b128ee1e-73c4-4428-ac6b-ea1c90161120 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.407301] env[62521]: DEBUG nova.compute.provider_tree [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.750103] env[62521]: INFO nova.compute.manager [-] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Took 1.02 seconds to deallocate network for instance. [ 731.753438] env[62521]: DEBUG nova.compute.claims [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 731.753607] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.803023] env[62521]: DEBUG oslo_concurrency.lockutils [req-85f3a95e-37eb-41db-a5ae-491ec60a518c req-e2ba22f1-11aa-435f-a3d5-3fb884eb93c5 service nova] Releasing lock "refresh_cache-5fa1c73e-3074-4fe0-8b4d-926bf02f1453" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.803023] env[62521]: DEBUG nova.compute.manager [req-85f3a95e-37eb-41db-a5ae-491ec60a518c req-e2ba22f1-11aa-435f-a3d5-3fb884eb93c5 service nova] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Received event network-vif-deleted-1f89168c-d105-46ac-9320-34456e056199 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.910134] env[62521]: DEBUG nova.scheduler.client.report [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.982567] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.015377] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.015622] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.015776] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.015961] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.016164] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.016316] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.016522] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.016729] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.016837] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.016997] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.017194] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.018036] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b232a0a-8826-4591-a0ce-fb093474690d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.025841] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0b75a9-f195-43ab-943d-4a1dc2f20a72 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.113411] env[62521]: ERROR nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 444157f5-a8b1-4bc7-891f-25e923aaf7c2, please check neutron logs for more information. [ 732.113411] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.113411] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.113411] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.113411] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.113411] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.113411] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.113411] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.113411] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.113411] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 732.113411] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.113411] env[62521]: ERROR nova.compute.manager raise self.value [ 732.113411] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.113411] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.113411] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.113411] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.113847] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.113847] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.113847] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 444157f5-a8b1-4bc7-891f-25e923aaf7c2, please check neutron logs for more information. [ 732.113847] env[62521]: ERROR nova.compute.manager [ 732.113847] env[62521]: Traceback (most recent call last): [ 732.113847] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.113847] env[62521]: listener.cb(fileno) [ 732.113847] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.113847] env[62521]: result = function(*args, **kwargs) [ 732.113847] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.113847] env[62521]: return func(*args, **kwargs) [ 732.113847] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.113847] env[62521]: raise e [ 732.113847] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.113847] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 732.113847] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.113847] env[62521]: created_port_ids = self._update_ports_for_instance( [ 732.113847] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.113847] env[62521]: with excutils.save_and_reraise_exception(): [ 732.113847] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.113847] env[62521]: self.force_reraise() [ 732.113847] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.113847] env[62521]: raise self.value [ 732.113847] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.113847] env[62521]: updated_port = self._update_port( [ 732.113847] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.113847] env[62521]: _ensure_no_port_binding_failure(port) [ 732.113847] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.113847] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.115062] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 444157f5-a8b1-4bc7-891f-25e923aaf7c2, please check neutron logs for more information. [ 732.115062] env[62521]: Removing descriptor: 15 [ 732.115062] env[62521]: ERROR nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 444157f5-a8b1-4bc7-891f-25e923aaf7c2, please check neutron logs for more information. [ 732.115062] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Traceback (most recent call last): [ 732.115062] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.115062] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] yield resources [ 732.115062] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.115062] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] self.driver.spawn(context, instance, image_meta, [ 732.115062] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.115062] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.115062] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.115062] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] vm_ref = self.build_virtual_machine(instance, [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] for vif in network_info: [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] return self._sync_wrapper(fn, *args, **kwargs) [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] self.wait() [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] self[:] = self._gt.wait() [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] return self._exit_event.wait() [ 732.115597] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] result = hub.switch() [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] return self.greenlet.switch() [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] result = function(*args, **kwargs) [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] return func(*args, **kwargs) [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] raise e [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] nwinfo = self.network_api.allocate_for_instance( [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.116150] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] created_port_ids = self._update_ports_for_instance( [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] with excutils.save_and_reraise_exception(): [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] self.force_reraise() [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] raise self.value [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] updated_port = self._update_port( [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] _ensure_no_port_binding_failure(port) [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.116724] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] raise exception.PortBindingFailed(port_id=port['id']) [ 732.117315] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] nova.exception.PortBindingFailed: Binding failed for port 444157f5-a8b1-4bc7-891f-25e923aaf7c2, please check neutron logs for more information. [ 732.117315] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] [ 732.117315] env[62521]: INFO nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Terminating instance [ 732.117315] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "refresh_cache-79b55308-ff33-4b2a-bca8-30a1c5cf5916" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.117315] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquired lock "refresh_cache-79b55308-ff33-4b2a-bca8-30a1c5cf5916" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.117315] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.414744] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.415348] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 732.418328] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.655s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.419677] env[62521]: INFO nova.compute.claims [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.633530] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.639041] env[62521]: DEBUG nova.compute.manager [req-8aa7bca0-8840-4faa-a3ae-929e210a5c11 req-d0ad51e6-2b39-430d-8deb-862a1d4aa0aa service nova] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Received event network-changed-444157f5-a8b1-4bc7-891f-25e923aaf7c2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.639264] env[62521]: DEBUG nova.compute.manager [req-8aa7bca0-8840-4faa-a3ae-929e210a5c11 req-d0ad51e6-2b39-430d-8deb-862a1d4aa0aa service nova] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Refreshing instance network info cache due to event network-changed-444157f5-a8b1-4bc7-891f-25e923aaf7c2. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 732.639465] env[62521]: DEBUG oslo_concurrency.lockutils [req-8aa7bca0-8840-4faa-a3ae-929e210a5c11 req-d0ad51e6-2b39-430d-8deb-862a1d4aa0aa service nova] Acquiring lock "refresh_cache-79b55308-ff33-4b2a-bca8-30a1c5cf5916" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.724141] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.924190] env[62521]: DEBUG nova.compute.utils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 732.928790] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 732.928974] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 733.007028] env[62521]: DEBUG nova.policy [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8c0ef26e3b5414aa6f7ad999051f076', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2a99691b0c9a407fb90fd95bf9232c48', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 733.227355] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Releasing lock "refresh_cache-79b55308-ff33-4b2a-bca8-30a1c5cf5916" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.228132] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.228132] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.228853] env[62521]: DEBUG oslo_concurrency.lockutils [req-8aa7bca0-8840-4faa-a3ae-929e210a5c11 req-d0ad51e6-2b39-430d-8deb-862a1d4aa0aa service nova] Acquired lock "refresh_cache-79b55308-ff33-4b2a-bca8-30a1c5cf5916" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.228853] env[62521]: DEBUG nova.network.neutron [req-8aa7bca0-8840-4faa-a3ae-929e210a5c11 req-d0ad51e6-2b39-430d-8deb-862a1d4aa0aa service nova] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Refreshing network info cache for port 444157f5-a8b1-4bc7-891f-25e923aaf7c2 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 733.229656] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf46547b-6117-45a8-b3f0-4df7c2cb7507 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.239673] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cc50b6-8441-4679-a716-d20d08439ffa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.261632] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 79b55308-ff33-4b2a-bca8-30a1c5cf5916 could not be found. [ 733.261852] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.262035] env[62521]: INFO nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Took 0.03 seconds to destroy the instance on the hypervisor. [ 733.262273] env[62521]: DEBUG oslo.service.loopingcall [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.262632] env[62521]: DEBUG nova.compute.manager [-] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.262726] env[62521]: DEBUG nova.network.neutron [-] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.278032] env[62521]: DEBUG nova.network.neutron [-] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.346124] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Successfully created port: a0de19a8-3874-4edd-8e27-08865689670f {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 733.430312] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 733.751406] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ae307e-f4da-4920-b99b-ec4ebeeba2f8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.755135] env[62521]: DEBUG nova.network.neutron [req-8aa7bca0-8840-4faa-a3ae-929e210a5c11 req-d0ad51e6-2b39-430d-8deb-862a1d4aa0aa service nova] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.761516] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14ea813-2c6d-4ccd-af21-f24b9e0dba26 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.792412] env[62521]: DEBUG nova.network.neutron [-] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.795243] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb4dd89-4309-4ba6-927c-78c9f90d7b3b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.804483] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcea6b9a-edb4-430b-84b7-4c5097bc3d83 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.816873] env[62521]: DEBUG nova.compute.provider_tree [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.863355] env[62521]: DEBUG nova.network.neutron [req-8aa7bca0-8840-4faa-a3ae-929e210a5c11 req-d0ad51e6-2b39-430d-8deb-862a1d4aa0aa service nova] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.301260] env[62521]: INFO nova.compute.manager [-] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Took 1.04 seconds to deallocate network for instance. [ 734.301852] env[62521]: DEBUG nova.compute.claims [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 734.303715] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.321896] env[62521]: DEBUG nova.scheduler.client.report [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.366579] env[62521]: DEBUG oslo_concurrency.lockutils [req-8aa7bca0-8840-4faa-a3ae-929e210a5c11 req-d0ad51e6-2b39-430d-8deb-862a1d4aa0aa service nova] Releasing lock "refresh_cache-79b55308-ff33-4b2a-bca8-30a1c5cf5916" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.366882] env[62521]: DEBUG nova.compute.manager [req-8aa7bca0-8840-4faa-a3ae-929e210a5c11 req-d0ad51e6-2b39-430d-8deb-862a1d4aa0aa service nova] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Received event network-vif-deleted-444157f5-a8b1-4bc7-891f-25e923aaf7c2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.435726] env[62521]: ERROR nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a0de19a8-3874-4edd-8e27-08865689670f, please check neutron logs for more information. [ 734.435726] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 734.435726] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.435726] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 734.435726] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.435726] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 734.435726] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.435726] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 734.435726] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.435726] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 734.435726] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.435726] env[62521]: ERROR nova.compute.manager raise self.value [ 734.435726] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.435726] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 734.435726] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.435726] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 734.436378] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.436378] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 734.436378] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a0de19a8-3874-4edd-8e27-08865689670f, please check neutron logs for more information. [ 734.436378] env[62521]: ERROR nova.compute.manager [ 734.436378] env[62521]: Traceback (most recent call last): [ 734.436378] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 734.436378] env[62521]: listener.cb(fileno) [ 734.436378] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.436378] env[62521]: result = function(*args, **kwargs) [ 734.436378] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.436378] env[62521]: return func(*args, **kwargs) [ 734.436378] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.436378] env[62521]: raise e [ 734.436378] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.436378] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 734.436378] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.436378] env[62521]: created_port_ids = self._update_ports_for_instance( [ 734.436378] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.436378] env[62521]: with excutils.save_and_reraise_exception(): [ 734.436378] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.436378] env[62521]: self.force_reraise() [ 734.436378] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.436378] env[62521]: raise self.value [ 734.436378] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.436378] env[62521]: updated_port = self._update_port( [ 734.436378] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.436378] env[62521]: _ensure_no_port_binding_failure(port) [ 734.436378] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.436378] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 734.437309] env[62521]: nova.exception.PortBindingFailed: Binding failed for port a0de19a8-3874-4edd-8e27-08865689670f, please check neutron logs for more information. [ 734.437309] env[62521]: Removing descriptor: 15 [ 734.443368] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 734.478134] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.478493] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.478701] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.478929] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.479140] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.479310] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.479563] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.479692] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.479861] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.480034] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.480218] env[62521]: DEBUG nova.virt.hardware [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.481212] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a180de7a-bced-48fd-9421-e7e6bf488fe6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.491311] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3efc76-05ea-43bf-be82-c04610b54d52 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.506117] env[62521]: ERROR nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a0de19a8-3874-4edd-8e27-08865689670f, please check neutron logs for more information. [ 734.506117] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Traceback (most recent call last): [ 734.506117] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 734.506117] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] yield resources [ 734.506117] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.506117] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] self.driver.spawn(context, instance, image_meta, [ 734.506117] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 734.506117] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.506117] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.506117] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] vm_ref = self.build_virtual_machine(instance, [ 734.506117] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] for vif in network_info: [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] return self._sync_wrapper(fn, *args, **kwargs) [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] self.wait() [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] self[:] = self._gt.wait() [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] return self._exit_event.wait() [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 734.506625] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] current.throw(*self._exc) [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] result = function(*args, **kwargs) [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] return func(*args, **kwargs) [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] raise e [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] nwinfo = self.network_api.allocate_for_instance( [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] created_port_ids = self._update_ports_for_instance( [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] with excutils.save_and_reraise_exception(): [ 734.507220] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] self.force_reraise() [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] raise self.value [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] updated_port = self._update_port( [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] _ensure_no_port_binding_failure(port) [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] raise exception.PortBindingFailed(port_id=port['id']) [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] nova.exception.PortBindingFailed: Binding failed for port a0de19a8-3874-4edd-8e27-08865689670f, please check neutron logs for more information. [ 734.507691] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] [ 734.507691] env[62521]: INFO nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Terminating instance [ 734.508803] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "refresh_cache-5adbe047-7055-428f-b72b-67023c68cc04" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.508968] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquired lock "refresh_cache-5adbe047-7055-428f-b72b-67023c68cc04" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.509155] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.668792] env[62521]: DEBUG nova.compute.manager [req-1e9e8765-b087-4b60-9741-00a185861d59 req-40c73e83-0837-4f84-b64d-89ec325b3183 service nova] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Received event network-changed-a0de19a8-3874-4edd-8e27-08865689670f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.669015] env[62521]: DEBUG nova.compute.manager [req-1e9e8765-b087-4b60-9741-00a185861d59 req-40c73e83-0837-4f84-b64d-89ec325b3183 service nova] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Refreshing instance network info cache due to event network-changed-a0de19a8-3874-4edd-8e27-08865689670f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 734.669192] env[62521]: DEBUG oslo_concurrency.lockutils [req-1e9e8765-b087-4b60-9741-00a185861d59 req-40c73e83-0837-4f84-b64d-89ec325b3183 service nova] Acquiring lock "refresh_cache-5adbe047-7055-428f-b72b-67023c68cc04" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.828029] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.828727] env[62521]: DEBUG nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 734.831415] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.085s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.833171] env[62521]: INFO nova.compute.claims [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.036577] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.159466] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.340259] env[62521]: DEBUG nova.compute.utils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.341570] env[62521]: DEBUG nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 735.341744] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 735.385062] env[62521]: DEBUG nova.policy [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '39a60744d425428c88a71cc29061a1cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '941c76a8bff8493288a960a452bb1343', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 735.662846] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Releasing lock "refresh_cache-5adbe047-7055-428f-b72b-67023c68cc04" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.663585] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 735.663886] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.665077] env[62521]: DEBUG oslo_concurrency.lockutils [req-1e9e8765-b087-4b60-9741-00a185861d59 req-40c73e83-0837-4f84-b64d-89ec325b3183 service nova] Acquired lock "refresh_cache-5adbe047-7055-428f-b72b-67023c68cc04" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.665077] env[62521]: DEBUG nova.network.neutron [req-1e9e8765-b087-4b60-9741-00a185861d59 req-40c73e83-0837-4f84-b64d-89ec325b3183 service nova] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Refreshing network info cache for port a0de19a8-3874-4edd-8e27-08865689670f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 735.666272] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-69b0968e-f59d-4e4a-a1e6-f8471ac60f86 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.679382] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37f7ca1-b4df-4f4d-aa8a-a275cc92096a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.701909] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5adbe047-7055-428f-b72b-67023c68cc04 could not be found. [ 735.702177] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.702371] env[62521]: INFO nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Took 0.04 seconds to destroy the instance on the hypervisor. [ 735.702615] env[62521]: DEBUG oslo.service.loopingcall [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.702870] env[62521]: DEBUG nova.compute.manager [-] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.702997] env[62521]: DEBUG nova.network.neutron [-] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.719631] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Successfully created port: 7a5b2040-8d99-4975-80df-f4b5f477e01e {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 735.722143] env[62521]: DEBUG nova.network.neutron [-] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.846824] env[62521]: DEBUG nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.198949] env[62521]: DEBUG nova.network.neutron [req-1e9e8765-b087-4b60-9741-00a185861d59 req-40c73e83-0837-4f84-b64d-89ec325b3183 service nova] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.225459] env[62521]: DEBUG nova.network.neutron [-] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.249843] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef753d1b-117e-48b7-b10d-545c6447c6b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.262977] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d297081-2710-4eaa-bc63-b42d102bb7e0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.319597] env[62521]: DEBUG nova.network.neutron [req-1e9e8765-b087-4b60-9741-00a185861d59 req-40c73e83-0837-4f84-b64d-89ec325b3183 service nova] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.321976] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f60e4b0-99f1-4ebc-82c4-205df37f3b55 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.334160] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5efd772-18b0-4f48-89ab-d8d940b444dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.363236] env[62521]: DEBUG nova.compute.provider_tree [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.695063] env[62521]: DEBUG nova.compute.manager [req-b4cdd8a2-8eea-419b-bcf1-a7cb4f850d57 req-74fd5a7a-65d2-4dbc-8849-c5e82268d0b1 service nova] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Received event network-changed-7a5b2040-8d99-4975-80df-f4b5f477e01e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.695287] env[62521]: DEBUG nova.compute.manager [req-b4cdd8a2-8eea-419b-bcf1-a7cb4f850d57 req-74fd5a7a-65d2-4dbc-8849-c5e82268d0b1 service nova] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Refreshing instance network info cache due to event network-changed-7a5b2040-8d99-4975-80df-f4b5f477e01e. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 736.695502] env[62521]: DEBUG oslo_concurrency.lockutils [req-b4cdd8a2-8eea-419b-bcf1-a7cb4f850d57 req-74fd5a7a-65d2-4dbc-8849-c5e82268d0b1 service nova] Acquiring lock "refresh_cache-72e3b843-c54a-4de7-ace9-ee25c5df14bd" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.695643] env[62521]: DEBUG oslo_concurrency.lockutils [req-b4cdd8a2-8eea-419b-bcf1-a7cb4f850d57 req-74fd5a7a-65d2-4dbc-8849-c5e82268d0b1 service nova] Acquired lock "refresh_cache-72e3b843-c54a-4de7-ace9-ee25c5df14bd" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.695801] env[62521]: DEBUG nova.network.neutron [req-b4cdd8a2-8eea-419b-bcf1-a7cb4f850d57 req-74fd5a7a-65d2-4dbc-8849-c5e82268d0b1 service nova] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Refreshing network info cache for port 7a5b2040-8d99-4975-80df-f4b5f477e01e {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 736.729783] env[62521]: INFO nova.compute.manager [-] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Took 1.03 seconds to deallocate network for instance. [ 736.732133] env[62521]: DEBUG nova.compute.claims [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.732321] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.827481] env[62521]: DEBUG oslo_concurrency.lockutils [req-1e9e8765-b087-4b60-9741-00a185861d59 req-40c73e83-0837-4f84-b64d-89ec325b3183 service nova] Releasing lock "refresh_cache-5adbe047-7055-428f-b72b-67023c68cc04" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.827752] env[62521]: DEBUG nova.compute.manager [req-1e9e8765-b087-4b60-9741-00a185861d59 req-40c73e83-0837-4f84-b64d-89ec325b3183 service nova] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Received event network-vif-deleted-a0de19a8-3874-4edd-8e27-08865689670f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.841695] env[62521]: ERROR nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7a5b2040-8d99-4975-80df-f4b5f477e01e, please check neutron logs for more information. [ 736.841695] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 736.841695] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.841695] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 736.841695] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.841695] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 736.841695] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.841695] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 736.841695] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.841695] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 736.841695] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.841695] env[62521]: ERROR nova.compute.manager raise self.value [ 736.841695] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.841695] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 736.841695] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.841695] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 736.842184] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.842184] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 736.842184] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7a5b2040-8d99-4975-80df-f4b5f477e01e, please check neutron logs for more information. [ 736.842184] env[62521]: ERROR nova.compute.manager [ 736.842184] env[62521]: Traceback (most recent call last): [ 736.842184] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 736.842184] env[62521]: listener.cb(fileno) [ 736.842184] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.842184] env[62521]: result = function(*args, **kwargs) [ 736.842184] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.842184] env[62521]: return func(*args, **kwargs) [ 736.842184] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.842184] env[62521]: raise e [ 736.842184] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.842184] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 736.842184] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.842184] env[62521]: created_port_ids = self._update_ports_for_instance( [ 736.842184] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.842184] env[62521]: with excutils.save_and_reraise_exception(): [ 736.842184] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.842184] env[62521]: self.force_reraise() [ 736.842184] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.842184] env[62521]: raise self.value [ 736.842184] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.842184] env[62521]: updated_port = self._update_port( [ 736.842184] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.842184] env[62521]: _ensure_no_port_binding_failure(port) [ 736.842184] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.842184] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 736.843060] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 7a5b2040-8d99-4975-80df-f4b5f477e01e, please check neutron logs for more information. [ 736.843060] env[62521]: Removing descriptor: 15 [ 736.866198] env[62521]: DEBUG nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 736.869697] env[62521]: DEBUG nova.scheduler.client.report [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.901642] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.901891] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.902081] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.902294] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.902440] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.902774] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.903019] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.903209] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.903389] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.903551] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.903721] env[62521]: DEBUG nova.virt.hardware [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.904799] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b558a79-83b3-43b3-adfa-dfa34b16b115 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.913904] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031ed0d8-4442-48cf-af28-38e620d0bb7e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.927435] env[62521]: ERROR nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7a5b2040-8d99-4975-80df-f4b5f477e01e, please check neutron logs for more information. [ 736.927435] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Traceback (most recent call last): [ 736.927435] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 736.927435] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] yield resources [ 736.927435] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.927435] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] self.driver.spawn(context, instance, image_meta, [ 736.927435] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 736.927435] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.927435] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.927435] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] vm_ref = self.build_virtual_machine(instance, [ 736.927435] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] for vif in network_info: [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] return self._sync_wrapper(fn, *args, **kwargs) [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] self.wait() [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] self[:] = self._gt.wait() [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] return self._exit_event.wait() [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 736.927841] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] current.throw(*self._exc) [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] result = function(*args, **kwargs) [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] return func(*args, **kwargs) [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] raise e [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] nwinfo = self.network_api.allocate_for_instance( [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] created_port_ids = self._update_ports_for_instance( [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] with excutils.save_and_reraise_exception(): [ 736.928280] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] self.force_reraise() [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] raise self.value [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] updated_port = self._update_port( [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] _ensure_no_port_binding_failure(port) [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] raise exception.PortBindingFailed(port_id=port['id']) [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] nova.exception.PortBindingFailed: Binding failed for port 7a5b2040-8d99-4975-80df-f4b5f477e01e, please check neutron logs for more information. [ 736.928681] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] [ 736.928681] env[62521]: INFO nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Terminating instance [ 736.929493] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "refresh_cache-72e3b843-c54a-4de7-ace9-ee25c5df14bd" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.213685] env[62521]: DEBUG nova.network.neutron [req-b4cdd8a2-8eea-419b-bcf1-a7cb4f850d57 req-74fd5a7a-65d2-4dbc-8849-c5e82268d0b1 service nova] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.325685] env[62521]: DEBUG nova.network.neutron [req-b4cdd8a2-8eea-419b-bcf1-a7cb4f850d57 req-74fd5a7a-65d2-4dbc-8849-c5e82268d0b1 service nova] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.374363] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.375156] env[62521]: DEBUG nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 737.377467] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.370s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.828054] env[62521]: DEBUG oslo_concurrency.lockutils [req-b4cdd8a2-8eea-419b-bcf1-a7cb4f850d57 req-74fd5a7a-65d2-4dbc-8849-c5e82268d0b1 service nova] Releasing lock "refresh_cache-72e3b843-c54a-4de7-ace9-ee25c5df14bd" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.828486] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquired lock "refresh_cache-72e3b843-c54a-4de7-ace9-ee25c5df14bd" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.829048] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.884300] env[62521]: DEBUG nova.compute.utils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 737.886411] env[62521]: DEBUG nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 737.886411] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 737.924442] env[62521]: DEBUG nova.policy [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e265e90b68a44f2ea47fd3828733758e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28071af632ef48bd87a2ccfe842c206d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.200406] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a7c08b-7086-40e2-9390-43b5dca6c023 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.208045] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270111ad-083b-49a5-a91a-9116ca6a4230 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.239523] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf716a1-d9c2-4e1f-8573-475fae6db391 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.246720] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6eca60-6186-4f85-9f5b-c6cdc5c47c25 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.259385] env[62521]: DEBUG nova.compute.provider_tree [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.346897] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.390680] env[62521]: DEBUG nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 738.434495] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.434495] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Successfully created port: b5049c47-da79-4cb5-979e-f6d73ac5c55a {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 738.762506] env[62521]: DEBUG nova.scheduler.client.report [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.807511] env[62521]: DEBUG nova.compute.manager [req-795ab187-139e-4f9c-bb3c-02ef31249d3e req-9a7bd9e4-4338-4801-80e1-dbe3d69c3906 service nova] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Received event network-vif-deleted-7a5b2040-8d99-4975-80df-f4b5f477e01e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.937049] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Releasing lock "refresh_cache-72e3b843-c54a-4de7-ace9-ee25c5df14bd" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.937573] env[62521]: DEBUG nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 738.938312] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 738.939517] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d62e8e24-2e5b-4cb1-90f5-7d52681c589b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.948988] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2aceb81-af84-4f18-b808-8d3fb5016278 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.972146] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 72e3b843-c54a-4de7-ace9-ee25c5df14bd could not be found. [ 738.972359] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 738.972534] env[62521]: INFO nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 738.972762] env[62521]: DEBUG oslo.service.loopingcall [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.972966] env[62521]: DEBUG nova.compute.manager [-] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.973084] env[62521]: DEBUG nova.network.neutron [-] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.987171] env[62521]: DEBUG nova.network.neutron [-] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.267400] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.890s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.268031] env[62521]: ERROR nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 967c9d3d-95b2-4703-b878-4787a02da1f5, please check neutron logs for more information. [ 739.268031] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Traceback (most recent call last): [ 739.268031] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.268031] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] self.driver.spawn(context, instance, image_meta, [ 739.268031] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 739.268031] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.268031] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.268031] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] vm_ref = self.build_virtual_machine(instance, [ 739.268031] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.268031] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.268031] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] for vif in network_info: [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] return self._sync_wrapper(fn, *args, **kwargs) [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] self.wait() [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] self[:] = self._gt.wait() [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] return self._exit_event.wait() [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] result = hub.switch() [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.268351] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] return self.greenlet.switch() [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] result = function(*args, **kwargs) [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] return func(*args, **kwargs) [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] raise e [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] nwinfo = self.network_api.allocate_for_instance( [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] created_port_ids = self._update_ports_for_instance( [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] with excutils.save_and_reraise_exception(): [ 739.268697] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] self.force_reraise() [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] raise self.value [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] updated_port = self._update_port( [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] _ensure_no_port_binding_failure(port) [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] raise exception.PortBindingFailed(port_id=port['id']) [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] nova.exception.PortBindingFailed: Binding failed for port 967c9d3d-95b2-4703-b878-4787a02da1f5, please check neutron logs for more information. [ 739.269048] env[62521]: ERROR nova.compute.manager [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] [ 739.269349] env[62521]: DEBUG nova.compute.utils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Binding failed for port 967c9d3d-95b2-4703-b878-4787a02da1f5, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.271570] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.599s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.274332] env[62521]: DEBUG nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Build of instance cf71efc9-cfa1-4856-a733-5434bf4465b1 was re-scheduled: Binding failed for port 967c9d3d-95b2-4703-b878-4787a02da1f5, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 739.274769] env[62521]: DEBUG nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 739.275040] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquiring lock "refresh_cache-cf71efc9-cfa1-4856-a733-5434bf4465b1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.275216] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Acquired lock "refresh_cache-cf71efc9-cfa1-4856-a733-5434bf4465b1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.275416] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.400851] env[62521]: DEBUG nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 739.426246] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.426498] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.426656] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.426833] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.426975] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.427137] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.427349] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.427514] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.428306] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.428306] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.428306] env[62521]: DEBUG nova.virt.hardware [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.428841] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875bb65c-db61-4dcd-ae72-ad0ac736cff3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.436816] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4051c5-85a7-4a8a-aba2-8b196dd2e35f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.486553] env[62521]: ERROR nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b5049c47-da79-4cb5-979e-f6d73ac5c55a, please check neutron logs for more information. [ 739.486553] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 739.486553] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.486553] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 739.486553] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.486553] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 739.486553] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.486553] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 739.486553] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.486553] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 739.486553] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.486553] env[62521]: ERROR nova.compute.manager raise self.value [ 739.486553] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.486553] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 739.486553] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.486553] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 739.487085] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.487085] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 739.487085] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b5049c47-da79-4cb5-979e-f6d73ac5c55a, please check neutron logs for more information. [ 739.487085] env[62521]: ERROR nova.compute.manager [ 739.487085] env[62521]: Traceback (most recent call last): [ 739.487085] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 739.487085] env[62521]: listener.cb(fileno) [ 739.487085] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.487085] env[62521]: result = function(*args, **kwargs) [ 739.487085] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.487085] env[62521]: return func(*args, **kwargs) [ 739.487085] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.487085] env[62521]: raise e [ 739.487085] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.487085] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 739.487085] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.487085] env[62521]: created_port_ids = self._update_ports_for_instance( [ 739.487085] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.487085] env[62521]: with excutils.save_and_reraise_exception(): [ 739.487085] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.487085] env[62521]: self.force_reraise() [ 739.487085] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.487085] env[62521]: raise self.value [ 739.487085] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.487085] env[62521]: updated_port = self._update_port( [ 739.487085] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.487085] env[62521]: _ensure_no_port_binding_failure(port) [ 739.487085] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.487085] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 739.487973] env[62521]: nova.exception.PortBindingFailed: Binding failed for port b5049c47-da79-4cb5-979e-f6d73ac5c55a, please check neutron logs for more information. [ 739.487973] env[62521]: Removing descriptor: 15 [ 739.487973] env[62521]: ERROR nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b5049c47-da79-4cb5-979e-f6d73ac5c55a, please check neutron logs for more information. [ 739.487973] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] Traceback (most recent call last): [ 739.487973] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 739.487973] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] yield resources [ 739.487973] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.487973] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] self.driver.spawn(context, instance, image_meta, [ 739.487973] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 739.487973] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.487973] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.487973] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] vm_ref = self.build_virtual_machine(instance, [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] for vif in network_info: [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] return self._sync_wrapper(fn, *args, **kwargs) [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] self.wait() [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] self[:] = self._gt.wait() [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] return self._exit_event.wait() [ 739.488380] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] result = hub.switch() [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] return self.greenlet.switch() [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] result = function(*args, **kwargs) [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] return func(*args, **kwargs) [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] raise e [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] nwinfo = self.network_api.allocate_for_instance( [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.488873] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] created_port_ids = self._update_ports_for_instance( [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] with excutils.save_and_reraise_exception(): [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] self.force_reraise() [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] raise self.value [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] updated_port = self._update_port( [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] _ensure_no_port_binding_failure(port) [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.489329] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] raise exception.PortBindingFailed(port_id=port['id']) [ 739.489842] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] nova.exception.PortBindingFailed: Binding failed for port b5049c47-da79-4cb5-979e-f6d73ac5c55a, please check neutron logs for more information. [ 739.489842] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] [ 739.489842] env[62521]: INFO nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Terminating instance [ 739.489842] env[62521]: DEBUG nova.network.neutron [-] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.490508] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquiring lock "refresh_cache-86727bdc-342a-429c-85ce-adec8823b799" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.490669] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquired lock "refresh_cache-86727bdc-342a-429c-85ce-adec8823b799" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.490829] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.802013] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.921998] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.992411] env[62521]: INFO nova.compute.manager [-] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Took 1.02 seconds to deallocate network for instance. [ 739.997695] env[62521]: DEBUG nova.compute.claims [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 739.997878] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.010302] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.078916] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.086246] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0c733c-529b-4286-aa0f-b3cba9a3fdb4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.094018] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fcb3dc-2412-4b44-8b49-79bbcb0bb53d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.124268] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9588f109-f2fe-47bf-bd0b-89f4219fc6e4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.130727] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a91398-3ed1-459b-8a79-b6932bc28397 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.143185] env[62521]: DEBUG nova.compute.provider_tree [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.426027] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Releasing lock "refresh_cache-cf71efc9-cfa1-4856-a733-5434bf4465b1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.426027] env[62521]: DEBUG nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 740.426027] env[62521]: DEBUG nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.426027] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.441261] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.581699] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Releasing lock "refresh_cache-86727bdc-342a-429c-85ce-adec8823b799" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.582146] env[62521]: DEBUG nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 740.582341] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 740.582622] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-53bc8da5-2266-44b0-be59-d2c3cd232704 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.592642] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4716e942-a340-4b0e-bca5-a52dec2d88e8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.612952] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 86727bdc-342a-429c-85ce-adec8823b799 could not be found. [ 740.613201] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.613384] env[62521]: INFO nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Took 0.03 seconds to destroy the instance on the hypervisor. [ 740.613621] env[62521]: DEBUG oslo.service.loopingcall [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.613829] env[62521]: DEBUG nova.compute.manager [-] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.613924] env[62521]: DEBUG nova.network.neutron [-] [instance: 86727bdc-342a-429c-85ce-adec8823b799] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.634704] env[62521]: DEBUG nova.network.neutron [-] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.646355] env[62521]: DEBUG nova.scheduler.client.report [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.830639] env[62521]: DEBUG nova.compute.manager [req-8a75f3b0-6737-4515-9cc1-834d1e30d20b req-a3d6f7b5-9c16-4778-8fca-592a479bccb5 service nova] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Received event network-changed-b5049c47-da79-4cb5-979e-f6d73ac5c55a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.830851] env[62521]: DEBUG nova.compute.manager [req-8a75f3b0-6737-4515-9cc1-834d1e30d20b req-a3d6f7b5-9c16-4778-8fca-592a479bccb5 service nova] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Refreshing instance network info cache due to event network-changed-b5049c47-da79-4cb5-979e-f6d73ac5c55a. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 740.831080] env[62521]: DEBUG oslo_concurrency.lockutils [req-8a75f3b0-6737-4515-9cc1-834d1e30d20b req-a3d6f7b5-9c16-4778-8fca-592a479bccb5 service nova] Acquiring lock "refresh_cache-86727bdc-342a-429c-85ce-adec8823b799" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.831227] env[62521]: DEBUG oslo_concurrency.lockutils [req-8a75f3b0-6737-4515-9cc1-834d1e30d20b req-a3d6f7b5-9c16-4778-8fca-592a479bccb5 service nova] Acquired lock "refresh_cache-86727bdc-342a-429c-85ce-adec8823b799" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.831390] env[62521]: DEBUG nova.network.neutron [req-8a75f3b0-6737-4515-9cc1-834d1e30d20b req-a3d6f7b5-9c16-4778-8fca-592a479bccb5 service nova] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Refreshing network info cache for port b5049c47-da79-4cb5-979e-f6d73ac5c55a {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 740.943580] env[62521]: DEBUG nova.network.neutron [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.137112] env[62521]: DEBUG nova.network.neutron [-] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.150609] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.879s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.151308] env[62521]: ERROR nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0e835e50-1f4e-4e89-b02e-5ef42824582f, please check neutron logs for more information. [ 741.151308] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Traceback (most recent call last): [ 741.151308] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.151308] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] self.driver.spawn(context, instance, image_meta, [ 741.151308] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 741.151308] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.151308] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.151308] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] vm_ref = self.build_virtual_machine(instance, [ 741.151308] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.151308] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.151308] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] for vif in network_info: [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] return self._sync_wrapper(fn, *args, **kwargs) [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] self.wait() [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] self[:] = self._gt.wait() [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] return self._exit_event.wait() [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] result = hub.switch() [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 741.151688] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] return self.greenlet.switch() [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] result = function(*args, **kwargs) [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] return func(*args, **kwargs) [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] raise e [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] nwinfo = self.network_api.allocate_for_instance( [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] created_port_ids = self._update_ports_for_instance( [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] with excutils.save_and_reraise_exception(): [ 741.152093] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] self.force_reraise() [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] raise self.value [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] updated_port = self._update_port( [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] _ensure_no_port_binding_failure(port) [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] raise exception.PortBindingFailed(port_id=port['id']) [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] nova.exception.PortBindingFailed: Binding failed for port 0e835e50-1f4e-4e89-b02e-5ef42824582f, please check neutron logs for more information. [ 741.152497] env[62521]: ERROR nova.compute.manager [instance: d7575b46-ecf9-4154-92b1-86119e727dda] [ 741.152841] env[62521]: DEBUG nova.compute.utils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Binding failed for port 0e835e50-1f4e-4e89-b02e-5ef42824582f, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 741.153104] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.303s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.154856] env[62521]: INFO nova.compute.claims [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.157557] env[62521]: DEBUG nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Build of instance d7575b46-ecf9-4154-92b1-86119e727dda was re-scheduled: Binding failed for port 0e835e50-1f4e-4e89-b02e-5ef42824582f, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 741.157968] env[62521]: DEBUG nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 741.158200] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Acquiring lock "refresh_cache-d7575b46-ecf9-4154-92b1-86119e727dda" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.158347] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Acquired lock "refresh_cache-d7575b46-ecf9-4154-92b1-86119e727dda" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.158505] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 741.346138] env[62521]: DEBUG nova.network.neutron [req-8a75f3b0-6737-4515-9cc1-834d1e30d20b req-a3d6f7b5-9c16-4778-8fca-592a479bccb5 service nova] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.410268] env[62521]: DEBUG nova.network.neutron [req-8a75f3b0-6737-4515-9cc1-834d1e30d20b req-a3d6f7b5-9c16-4778-8fca-592a479bccb5 service nova] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.446207] env[62521]: INFO nova.compute.manager [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] [instance: cf71efc9-cfa1-4856-a733-5434bf4465b1] Took 1.02 seconds to deallocate network for instance. [ 741.639619] env[62521]: INFO nova.compute.manager [-] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Took 1.03 seconds to deallocate network for instance. [ 741.642198] env[62521]: DEBUG nova.compute.claims [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 741.642383] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.682033] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.793451] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.915641] env[62521]: DEBUG oslo_concurrency.lockutils [req-8a75f3b0-6737-4515-9cc1-834d1e30d20b req-a3d6f7b5-9c16-4778-8fca-592a479bccb5 service nova] Releasing lock "refresh_cache-86727bdc-342a-429c-85ce-adec8823b799" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.915915] env[62521]: DEBUG nova.compute.manager [req-8a75f3b0-6737-4515-9cc1-834d1e30d20b req-a3d6f7b5-9c16-4778-8fca-592a479bccb5 service nova] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Received event network-vif-deleted-b5049c47-da79-4cb5-979e-f6d73ac5c55a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.296026] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Releasing lock "refresh_cache-d7575b46-ecf9-4154-92b1-86119e727dda" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.296135] env[62521]: DEBUG nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 742.296274] env[62521]: DEBUG nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.296448] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 742.311864] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.470533] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe4dff5-166d-4f4f-ad5a-e6898925c563 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.473657] env[62521]: INFO nova.scheduler.client.report [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Deleted allocations for instance cf71efc9-cfa1-4856-a733-5434bf4465b1 [ 742.485768] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e231c77f-670e-4774-885b-23c0d966142a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.516351] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546df009-c195-4644-9988-63ad84b87668 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.524017] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5934125e-dcbc-4150-baaf-bc552847caa8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.541670] env[62521]: DEBUG nova.compute.provider_tree [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.816148] env[62521]: DEBUG nova.network.neutron [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.990282] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9a6ab324-4704-491f-8206-03136ff4a49f tempest-VolumesAdminNegativeTest-669787752 tempest-VolumesAdminNegativeTest-669787752-project-member] Lock "cf71efc9-cfa1-4856-a733-5434bf4465b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.134s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.044839] env[62521]: DEBUG nova.scheduler.client.report [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.318496] env[62521]: INFO nova.compute.manager [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] [instance: d7575b46-ecf9-4154-92b1-86119e727dda] Took 1.02 seconds to deallocate network for instance. [ 743.493254] env[62521]: DEBUG nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.554743] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.555340] env[62521]: DEBUG nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 743.557826] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.756s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.017296] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.064028] env[62521]: DEBUG nova.compute.utils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.068678] env[62521]: DEBUG nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.069038] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 744.129766] env[62521]: DEBUG nova.policy [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0880cde418fd4cf1b03c7c2fe2a3991c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e6034d8bc2645e491d404fd8ba25c3c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 744.348548] env[62521]: INFO nova.scheduler.client.report [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Deleted allocations for instance d7575b46-ecf9-4154-92b1-86119e727dda [ 744.494134] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf54e54-6c10-4936-bd34-80dd18d73b57 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.503986] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d912ab19-295a-4a30-93b8-cd73174866d1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.536542] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8263a2-ecc3-43b9-bb16-04f1bbb5f4f6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.543086] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5f51e3-49f4-4483-b9ae-dc1f748ab709 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.557307] env[62521]: DEBUG nova.compute.provider_tree [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.568916] env[62521]: DEBUG nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 744.571741] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Successfully created port: d72099b7-8aec-49f1-b2c8-c1b080e3f10d {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 744.863603] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32bfd0dd-f364-406b-81ac-a2e72051b6e6 tempest-ServerAddressesTestJSON-621897147 tempest-ServerAddressesTestJSON-621897147-project-member] Lock "d7575b46-ecf9-4154-92b1-86119e727dda" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.245s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.060878] env[62521]: DEBUG nova.scheduler.client.report [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.366169] env[62521]: DEBUG nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 745.566760] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.009s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.567660] env[62521]: ERROR nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c621baa7-5581-4e73-a0df-0b154224d696, please check neutron logs for more information. [ 745.567660] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Traceback (most recent call last): [ 745.567660] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.567660] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] self.driver.spawn(context, instance, image_meta, [ 745.567660] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 745.567660] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.567660] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.567660] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] vm_ref = self.build_virtual_machine(instance, [ 745.567660] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.567660] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.567660] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] for vif in network_info: [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] return self._sync_wrapper(fn, *args, **kwargs) [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] self.wait() [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] self[:] = self._gt.wait() [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] return self._exit_event.wait() [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] result = hub.switch() [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 745.567986] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] return self.greenlet.switch() [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] result = function(*args, **kwargs) [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] return func(*args, **kwargs) [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] raise e [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] nwinfo = self.network_api.allocate_for_instance( [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] created_port_ids = self._update_ports_for_instance( [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] with excutils.save_and_reraise_exception(): [ 745.568364] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] self.force_reraise() [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] raise self.value [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] updated_port = self._update_port( [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] _ensure_no_port_binding_failure(port) [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] raise exception.PortBindingFailed(port_id=port['id']) [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] nova.exception.PortBindingFailed: Binding failed for port c621baa7-5581-4e73-a0df-0b154224d696, please check neutron logs for more information. [ 745.568713] env[62521]: ERROR nova.compute.manager [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] [ 745.569029] env[62521]: DEBUG nova.compute.utils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Binding failed for port c621baa7-5581-4e73-a0df-0b154224d696, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.569877] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.273s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.573483] env[62521]: DEBUG nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Build of instance 0d72d988-f96b-428c-89bf-a96572f71db5 was re-scheduled: Binding failed for port c621baa7-5581-4e73-a0df-0b154224d696, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.573944] env[62521]: DEBUG nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.574185] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "refresh_cache-0d72d988-f96b-428c-89bf-a96572f71db5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.574333] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquired lock "refresh_cache-0d72d988-f96b-428c-89bf-a96572f71db5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.574493] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.580906] env[62521]: DEBUG nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.617803] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.618056] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.618222] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.619865] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.619865] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.623027] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.623027] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.623027] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.623344] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.623344] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.623477] env[62521]: DEBUG nova.virt.hardware [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.625143] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202fd6b8-b4c3-4a4f-a1bd-f676dcaec0de {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.635477] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12821a9-d4cf-4bf0-a6f2-ecc559a52112 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.812657] env[62521]: DEBUG nova.compute.manager [req-1520b06a-803b-4e94-98f6-f9fdb77a83a0 req-0bc051c9-7f89-45c0-bba4-20ef20c9e712 service nova] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Received event network-changed-d72099b7-8aec-49f1-b2c8-c1b080e3f10d {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 745.812864] env[62521]: DEBUG nova.compute.manager [req-1520b06a-803b-4e94-98f6-f9fdb77a83a0 req-0bc051c9-7f89-45c0-bba4-20ef20c9e712 service nova] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Refreshing instance network info cache due to event network-changed-d72099b7-8aec-49f1-b2c8-c1b080e3f10d. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 745.813095] env[62521]: DEBUG oslo_concurrency.lockutils [req-1520b06a-803b-4e94-98f6-f9fdb77a83a0 req-0bc051c9-7f89-45c0-bba4-20ef20c9e712 service nova] Acquiring lock "refresh_cache-277ecbd7-e770-4ea0-ae2e-b49598428d0f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.813240] env[62521]: DEBUG oslo_concurrency.lockutils [req-1520b06a-803b-4e94-98f6-f9fdb77a83a0 req-0bc051c9-7f89-45c0-bba4-20ef20c9e712 service nova] Acquired lock "refresh_cache-277ecbd7-e770-4ea0-ae2e-b49598428d0f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.813399] env[62521]: DEBUG nova.network.neutron [req-1520b06a-803b-4e94-98f6-f9fdb77a83a0 req-0bc051c9-7f89-45c0-bba4-20ef20c9e712 service nova] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Refreshing network info cache for port d72099b7-8aec-49f1-b2c8-c1b080e3f10d {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 745.894764] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.056235] env[62521]: ERROR nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d72099b7-8aec-49f1-b2c8-c1b080e3f10d, please check neutron logs for more information. [ 746.056235] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.056235] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.056235] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.056235] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.056235] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.056235] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.056235] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.056235] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.056235] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 746.056235] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.056235] env[62521]: ERROR nova.compute.manager raise self.value [ 746.056235] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.056235] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.056235] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.056235] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.056775] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.056775] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.056775] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d72099b7-8aec-49f1-b2c8-c1b080e3f10d, please check neutron logs for more information. [ 746.056775] env[62521]: ERROR nova.compute.manager [ 746.056775] env[62521]: Traceback (most recent call last): [ 746.056775] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.056775] env[62521]: listener.cb(fileno) [ 746.056775] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.056775] env[62521]: result = function(*args, **kwargs) [ 746.056775] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.056775] env[62521]: return func(*args, **kwargs) [ 746.056775] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.056775] env[62521]: raise e [ 746.056775] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.056775] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 746.056775] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.056775] env[62521]: created_port_ids = self._update_ports_for_instance( [ 746.056775] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.056775] env[62521]: with excutils.save_and_reraise_exception(): [ 746.056775] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.056775] env[62521]: self.force_reraise() [ 746.056775] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.056775] env[62521]: raise self.value [ 746.056775] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.056775] env[62521]: updated_port = self._update_port( [ 746.056775] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.056775] env[62521]: _ensure_no_port_binding_failure(port) [ 746.056775] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.056775] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.057609] env[62521]: nova.exception.PortBindingFailed: Binding failed for port d72099b7-8aec-49f1-b2c8-c1b080e3f10d, please check neutron logs for more information. [ 746.057609] env[62521]: Removing descriptor: 15 [ 746.057609] env[62521]: ERROR nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d72099b7-8aec-49f1-b2c8-c1b080e3f10d, please check neutron logs for more information. [ 746.057609] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Traceback (most recent call last): [ 746.057609] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 746.057609] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] yield resources [ 746.057609] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.057609] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] self.driver.spawn(context, instance, image_meta, [ 746.057609] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 746.057609] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.057609] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.057609] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] vm_ref = self.build_virtual_machine(instance, [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] for vif in network_info: [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] return self._sync_wrapper(fn, *args, **kwargs) [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] self.wait() [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] self[:] = self._gt.wait() [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] return self._exit_event.wait() [ 746.058046] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] result = hub.switch() [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] return self.greenlet.switch() [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] result = function(*args, **kwargs) [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] return func(*args, **kwargs) [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] raise e [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] nwinfo = self.network_api.allocate_for_instance( [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.058424] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] created_port_ids = self._update_ports_for_instance( [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] with excutils.save_and_reraise_exception(): [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] self.force_reraise() [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] raise self.value [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] updated_port = self._update_port( [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] _ensure_no_port_binding_failure(port) [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.058788] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] raise exception.PortBindingFailed(port_id=port['id']) [ 746.059150] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] nova.exception.PortBindingFailed: Binding failed for port d72099b7-8aec-49f1-b2c8-c1b080e3f10d, please check neutron logs for more information. [ 746.059150] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] [ 746.059150] env[62521]: INFO nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Terminating instance [ 746.061336] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Acquiring lock "refresh_cache-277ecbd7-e770-4ea0-ae2e-b49598428d0f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.108735] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.307678] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.345162] env[62521]: DEBUG nova.network.neutron [req-1520b06a-803b-4e94-98f6-f9fdb77a83a0 req-0bc051c9-7f89-45c0-bba4-20ef20c9e712 service nova] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.431762] env[62521]: DEBUG nova.network.neutron [req-1520b06a-803b-4e94-98f6-f9fdb77a83a0 req-0bc051c9-7f89-45c0-bba4-20ef20c9e712 service nova] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.499180] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658e1b12-af16-45ce-a633-6b78c3b6957a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.508246] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ede34f6-07db-43d5-becc-2ab0a6aaadd5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.539502] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5fed850-e7a7-4bb7-a076-2e15148f0062 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.546983] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec0376f-e873-4fd3-aa72-3bcd460805ac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.559922] env[62521]: DEBUG nova.compute.provider_tree [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.814390] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Releasing lock "refresh_cache-0d72d988-f96b-428c-89bf-a96572f71db5" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.814684] env[62521]: DEBUG nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 746.814787] env[62521]: DEBUG nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.814948] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.842227] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.940398] env[62521]: DEBUG oslo_concurrency.lockutils [req-1520b06a-803b-4e94-98f6-f9fdb77a83a0 req-0bc051c9-7f89-45c0-bba4-20ef20c9e712 service nova] Releasing lock "refresh_cache-277ecbd7-e770-4ea0-ae2e-b49598428d0f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.940398] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Acquired lock "refresh_cache-277ecbd7-e770-4ea0-ae2e-b49598428d0f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.940398] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.062859] env[62521]: DEBUG nova.scheduler.client.report [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.346581] env[62521]: DEBUG nova.network.neutron [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.465538] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.545254] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.568632] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.998s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.569071] env[62521]: ERROR nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6a4810e6-a03d-46be-bc14-9448caf0aebb, please check neutron logs for more information. [ 747.569071] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Traceback (most recent call last): [ 747.569071] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.569071] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] self.driver.spawn(context, instance, image_meta, [ 747.569071] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 747.569071] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.569071] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.569071] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] vm_ref = self.build_virtual_machine(instance, [ 747.569071] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.569071] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.569071] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] for vif in network_info: [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] return self._sync_wrapper(fn, *args, **kwargs) [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] self.wait() [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] self[:] = self._gt.wait() [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] return self._exit_event.wait() [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] result = hub.switch() [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 747.569493] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] return self.greenlet.switch() [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] result = function(*args, **kwargs) [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] return func(*args, **kwargs) [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] raise e [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] nwinfo = self.network_api.allocate_for_instance( [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] created_port_ids = self._update_ports_for_instance( [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] with excutils.save_and_reraise_exception(): [ 747.569851] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] self.force_reraise() [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] raise self.value [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] updated_port = self._update_port( [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] _ensure_no_port_binding_failure(port) [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] raise exception.PortBindingFailed(port_id=port['id']) [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] nova.exception.PortBindingFailed: Binding failed for port 6a4810e6-a03d-46be-bc14-9448caf0aebb, please check neutron logs for more information. [ 747.570210] env[62521]: ERROR nova.compute.manager [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] [ 747.570507] env[62521]: DEBUG nova.compute.utils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Binding failed for port 6a4810e6-a03d-46be-bc14-9448caf0aebb, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 747.571248] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.818s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.573991] env[62521]: DEBUG nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Build of instance 884d2f4f-c0a2-4d08-962e-d6347dc9f678 was re-scheduled: Binding failed for port 6a4810e6-a03d-46be-bc14-9448caf0aebb, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 747.574431] env[62521]: DEBUG nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 747.574658] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "refresh_cache-884d2f4f-c0a2-4d08-962e-d6347dc9f678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.574815] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquired lock "refresh_cache-884d2f4f-c0a2-4d08-962e-d6347dc9f678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.574972] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.848897] env[62521]: INFO nova.compute.manager [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 0d72d988-f96b-428c-89bf-a96572f71db5] Took 1.03 seconds to deallocate network for instance. [ 747.853938] env[62521]: DEBUG nova.compute.manager [req-98db3785-ba87-415a-a51a-5467d8e1a962 req-5a88f29a-1b7d-44a9-923f-f5341f72f108 service nova] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Received event network-vif-deleted-d72099b7-8aec-49f1-b2c8-c1b080e3f10d {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.048602] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Releasing lock "refresh_cache-277ecbd7-e770-4ea0-ae2e-b49598428d0f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.048664] env[62521]: DEBUG nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 748.048852] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 748.049174] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc121912-a9e6-4034-a6ce-487ac8594282 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.057857] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b06e0d7-35c6-45fc-b77e-9e398ea73e30 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.084374] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 277ecbd7-e770-4ea0-ae2e-b49598428d0f could not be found. [ 748.084585] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.084759] env[62521]: INFO nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 748.084994] env[62521]: DEBUG oslo.service.loopingcall [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.085500] env[62521]: DEBUG nova.compute.manager [-] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.085632] env[62521]: DEBUG nova.network.neutron [-] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.104920] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.106273] env[62521]: DEBUG nova.network.neutron [-] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.234232] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.466391] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f62b61d-06f5-4ad7-8a3f-800fbee27450 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.474051] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2c3edd-1a07-4d64-b055-fd0e2b21a508 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.504923] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb8bffd-b414-4b56-8c43-2c75511241ec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.511930] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d92efc-d7d9-4bec-a6df-50f8a9e316e3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.524819] env[62521]: DEBUG nova.compute.provider_tree [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.610058] env[62521]: DEBUG nova.network.neutron [-] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.736918] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Releasing lock "refresh_cache-884d2f4f-c0a2-4d08-962e-d6347dc9f678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.736918] env[62521]: DEBUG nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 748.737220] env[62521]: DEBUG nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.737467] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.754693] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.882237] env[62521]: INFO nova.scheduler.client.report [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Deleted allocations for instance 0d72d988-f96b-428c-89bf-a96572f71db5 [ 749.027759] env[62521]: DEBUG nova.scheduler.client.report [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.111635] env[62521]: INFO nova.compute.manager [-] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Took 1.03 seconds to deallocate network for instance. [ 749.114351] env[62521]: DEBUG nova.compute.claims [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 749.114569] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.259086] env[62521]: DEBUG nova.network.neutron [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.389768] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3c0f67ae-4405-44f1-b37b-f5a89893002a tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "0d72d988-f96b-428c-89bf-a96572f71db5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.676s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.533200] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.536395] env[62521]: ERROR nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f89168c-d105-46ac-9320-34456e056199, please check neutron logs for more information. [ 749.536395] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Traceback (most recent call last): [ 749.536395] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.536395] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] self.driver.spawn(context, instance, image_meta, [ 749.536395] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 749.536395] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.536395] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.536395] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] vm_ref = self.build_virtual_machine(instance, [ 749.536395] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.536395] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.536395] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] for vif in network_info: [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] return self._sync_wrapper(fn, *args, **kwargs) [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] self.wait() [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] self[:] = self._gt.wait() [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] return self._exit_event.wait() [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] current.throw(*self._exc) [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.536775] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] result = function(*args, **kwargs) [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] return func(*args, **kwargs) [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] raise e [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] nwinfo = self.network_api.allocate_for_instance( [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] created_port_ids = self._update_ports_for_instance( [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] with excutils.save_and_reraise_exception(): [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] self.force_reraise() [ 749.537137] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.537488] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] raise self.value [ 749.537488] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.537488] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] updated_port = self._update_port( [ 749.537488] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.537488] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] _ensure_no_port_binding_failure(port) [ 749.537488] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.537488] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] raise exception.PortBindingFailed(port_id=port['id']) [ 749.537488] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] nova.exception.PortBindingFailed: Binding failed for port 1f89168c-d105-46ac-9320-34456e056199, please check neutron logs for more information. [ 749.537488] env[62521]: ERROR nova.compute.manager [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] [ 749.537488] env[62521]: DEBUG nova.compute.utils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Binding failed for port 1f89168c-d105-46ac-9320-34456e056199, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.537762] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.234s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.542095] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Build of instance 5fa1c73e-3074-4fe0-8b4d-926bf02f1453 was re-scheduled: Binding failed for port 1f89168c-d105-46ac-9320-34456e056199, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 749.542547] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 749.542810] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "refresh_cache-5fa1c73e-3074-4fe0-8b4d-926bf02f1453" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.542912] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquired lock "refresh_cache-5fa1c73e-3074-4fe0-8b4d-926bf02f1453" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.543091] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.762170] env[62521]: INFO nova.compute.manager [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 884d2f4f-c0a2-4d08-962e-d6347dc9f678] Took 1.02 seconds to deallocate network for instance. [ 749.892946] env[62521]: DEBUG nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.066060] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.231298] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.399941] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb3a47b-2681-4e48-b493-5dfbf6600594 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.409297] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32d8f68-982a-4514-bc4b-020384abd480 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.413013] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.439125] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea80ee1-1523-4d63-bf28-ce39cc698082 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.446907] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0897485e-0234-4f92-931e-8ca8d92c7c6f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.460488] env[62521]: DEBUG nova.compute.provider_tree [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.732206] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Releasing lock "refresh_cache-5fa1c73e-3074-4fe0-8b4d-926bf02f1453" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.732829] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 750.732829] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.732829] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.751834] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.792409] env[62521]: INFO nova.scheduler.client.report [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Deleted allocations for instance 884d2f4f-c0a2-4d08-962e-d6347dc9f678 [ 750.964950] env[62521]: DEBUG nova.scheduler.client.report [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.254815] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.303917] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edf4794c-a3a1-4d72-9f7f-25c60d0646b9 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "884d2f4f-c0a2-4d08-962e-d6347dc9f678" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.721s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.469905] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.470569] env[62521]: ERROR nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 444157f5-a8b1-4bc7-891f-25e923aaf7c2, please check neutron logs for more information. [ 751.470569] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Traceback (most recent call last): [ 751.470569] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 751.470569] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] self.driver.spawn(context, instance, image_meta, [ 751.470569] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 751.470569] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.470569] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.470569] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] vm_ref = self.build_virtual_machine(instance, [ 751.470569] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.470569] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.470569] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] for vif in network_info: [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] return self._sync_wrapper(fn, *args, **kwargs) [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] self.wait() [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] self[:] = self._gt.wait() [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] return self._exit_event.wait() [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] result = hub.switch() [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 751.471174] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] return self.greenlet.switch() [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] result = function(*args, **kwargs) [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] return func(*args, **kwargs) [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] raise e [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] nwinfo = self.network_api.allocate_for_instance( [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] created_port_ids = self._update_ports_for_instance( [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] with excutils.save_and_reraise_exception(): [ 751.471714] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] self.force_reraise() [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] raise self.value [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] updated_port = self._update_port( [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] _ensure_no_port_binding_failure(port) [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] raise exception.PortBindingFailed(port_id=port['id']) [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] nova.exception.PortBindingFailed: Binding failed for port 444157f5-a8b1-4bc7-891f-25e923aaf7c2, please check neutron logs for more information. [ 751.472360] env[62521]: ERROR nova.compute.manager [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] [ 751.472846] env[62521]: DEBUG nova.compute.utils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Binding failed for port 444157f5-a8b1-4bc7-891f-25e923aaf7c2, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.472846] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.740s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.475977] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Build of instance 79b55308-ff33-4b2a-bca8-30a1c5cf5916 was re-scheduled: Binding failed for port 444157f5-a8b1-4bc7-891f-25e923aaf7c2, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 751.476503] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 751.477153] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "refresh_cache-79b55308-ff33-4b2a-bca8-30a1c5cf5916" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.477153] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquired lock "refresh_cache-79b55308-ff33-4b2a-bca8-30a1c5cf5916" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.477153] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.759047] env[62521]: INFO nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5fa1c73e-3074-4fe0-8b4d-926bf02f1453] Took 1.02 seconds to deallocate network for instance. [ 751.808429] env[62521]: DEBUG nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.000800] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.097544] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.330424] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.348318] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f5f11a-1286-4af6-9619-257d9dc6306c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.359364] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3e8ee3-c9ba-481b-b9e9-ddd33947aaea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.410032] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821f37a1-4713-4a0a-80d7-acdf041723c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.416122] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bff66fe-f3a3-43df-987a-ebb68830e35a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.430140] env[62521]: DEBUG nova.compute.provider_tree [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.601379] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Releasing lock "refresh_cache-79b55308-ff33-4b2a-bca8-30a1c5cf5916" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.601753] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 752.601954] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.602142] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.622131] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.793171] env[62521]: INFO nova.scheduler.client.report [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Deleted allocations for instance 5fa1c73e-3074-4fe0-8b4d-926bf02f1453 [ 752.933744] env[62521]: DEBUG nova.scheduler.client.report [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.124615] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.303897] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "5fa1c73e-3074-4fe0-8b4d-926bf02f1453" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.718s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.441994] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.441994] env[62521]: ERROR nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a0de19a8-3874-4edd-8e27-08865689670f, please check neutron logs for more information. [ 753.441994] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Traceback (most recent call last): [ 753.441994] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.441994] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] self.driver.spawn(context, instance, image_meta, [ 753.441994] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 753.441994] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.441994] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.441994] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] vm_ref = self.build_virtual_machine(instance, [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] for vif in network_info: [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] return self._sync_wrapper(fn, *args, **kwargs) [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] self.wait() [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] self[:] = self._gt.wait() [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] return self._exit_event.wait() [ 753.442521] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] current.throw(*self._exc) [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] result = function(*args, **kwargs) [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] return func(*args, **kwargs) [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] raise e [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] nwinfo = self.network_api.allocate_for_instance( [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] created_port_ids = self._update_ports_for_instance( [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.442926] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] with excutils.save_and_reraise_exception(): [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] self.force_reraise() [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] raise self.value [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] updated_port = self._update_port( [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] _ensure_no_port_binding_failure(port) [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] raise exception.PortBindingFailed(port_id=port['id']) [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] nova.exception.PortBindingFailed: Binding failed for port a0de19a8-3874-4edd-8e27-08865689670f, please check neutron logs for more information. [ 753.443366] env[62521]: ERROR nova.compute.manager [instance: 5adbe047-7055-428f-b72b-67023c68cc04] [ 753.443739] env[62521]: DEBUG nova.compute.utils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Binding failed for port a0de19a8-3874-4edd-8e27-08865689670f, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 753.443739] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.444s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.453075] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Build of instance 5adbe047-7055-428f-b72b-67023c68cc04 was re-scheduled: Binding failed for port a0de19a8-3874-4edd-8e27-08865689670f, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 753.453075] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 753.453075] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquiring lock "refresh_cache-5adbe047-7055-428f-b72b-67023c68cc04" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.453075] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Acquired lock "refresh_cache-5adbe047-7055-428f-b72b-67023c68cc04" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.453527] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.630168] env[62521]: INFO nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 79b55308-ff33-4b2a-bca8-30a1c5cf5916] Took 1.03 seconds to deallocate network for instance. [ 753.686592] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.686592] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.806906] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.973837] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.089373] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.284235] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ce5e16-920e-467b-b2c9-99be31a7ded9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.293233] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042a3ba2-335a-489d-b0ab-e6b51dbb6ad8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.328325] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7933b77c-4b7a-4084-a03d-90d3d58051a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.337406] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ee798c-e834-4790-b823-c7e0c96e734d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.344225] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.351327] env[62521]: DEBUG nova.compute.provider_tree [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.592402] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Releasing lock "refresh_cache-5adbe047-7055-428f-b72b-67023c68cc04" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.593643] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 754.593643] env[62521]: DEBUG nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.593771] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.616632] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.656031] env[62521]: INFO nova.scheduler.client.report [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Deleted allocations for instance 79b55308-ff33-4b2a-bca8-30a1c5cf5916 [ 754.854916] env[62521]: DEBUG nova.scheduler.client.report [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.120037] env[62521]: DEBUG nova.network.neutron [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.162911] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "79b55308-ff33-4b2a-bca8-30a1c5cf5916" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.544s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.362070] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.919s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.362070] env[62521]: ERROR nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7a5b2040-8d99-4975-80df-f4b5f477e01e, please check neutron logs for more information. [ 755.362070] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Traceback (most recent call last): [ 755.362070] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 755.362070] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] self.driver.spawn(context, instance, image_meta, [ 755.362070] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 755.362070] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.362070] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.362070] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] vm_ref = self.build_virtual_machine(instance, [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] for vif in network_info: [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] return self._sync_wrapper(fn, *args, **kwargs) [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] self.wait() [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] self[:] = self._gt.wait() [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] return self._exit_event.wait() [ 755.362651] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] current.throw(*self._exc) [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] result = function(*args, **kwargs) [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] return func(*args, **kwargs) [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] raise e [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] nwinfo = self.network_api.allocate_for_instance( [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] created_port_ids = self._update_ports_for_instance( [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.363086] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] with excutils.save_and_reraise_exception(): [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] self.force_reraise() [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] raise self.value [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] updated_port = self._update_port( [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] _ensure_no_port_binding_failure(port) [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] raise exception.PortBindingFailed(port_id=port['id']) [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] nova.exception.PortBindingFailed: Binding failed for port 7a5b2040-8d99-4975-80df-f4b5f477e01e, please check neutron logs for more information. [ 755.363648] env[62521]: ERROR nova.compute.manager [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] [ 755.364039] env[62521]: DEBUG nova.compute.utils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Binding failed for port 7a5b2040-8d99-4975-80df-f4b5f477e01e, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 755.364039] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.721s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.366063] env[62521]: DEBUG nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Build of instance 72e3b843-c54a-4de7-ace9-ee25c5df14bd was re-scheduled: Binding failed for port 7a5b2040-8d99-4975-80df-f4b5f477e01e, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 755.366536] env[62521]: DEBUG nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 755.366707] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquiring lock "refresh_cache-72e3b843-c54a-4de7-ace9-ee25c5df14bd" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.366852] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Acquired lock "refresh_cache-72e3b843-c54a-4de7-ace9-ee25c5df14bd" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.367024] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.623702] env[62521]: INFO nova.compute.manager [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] [instance: 5adbe047-7055-428f-b72b-67023c68cc04] Took 1.03 seconds to deallocate network for instance. [ 755.668992] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.889722] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.997418] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.179669] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1929df-0804-40c7-ac43-9157831e34a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.187676] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cce4259-3024-446e-bc81-7c84d9de9acc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.192818] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.218616] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae113c7d-f3b7-4062-b2e7-74c2d927d679 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.227197] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c640e7e-8a1f-4fe6-8dbf-6e7aee3aa117 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.241830] env[62521]: DEBUG nova.compute.provider_tree [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.500492] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Releasing lock "refresh_cache-72e3b843-c54a-4de7-ace9-ee25c5df14bd" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.500766] env[62521]: DEBUG nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 756.500856] env[62521]: DEBUG nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.501622] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 756.515020] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.653898] env[62521]: INFO nova.scheduler.client.report [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Deleted allocations for instance 5adbe047-7055-428f-b72b-67023c68cc04 [ 756.745500] env[62521]: DEBUG nova.scheduler.client.report [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.017934] env[62521]: DEBUG nova.network.neutron [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.161674] env[62521]: DEBUG oslo_concurrency.lockutils [None req-100eb175-35dd-44d4-bae9-0e976aca4cf7 tempest-ListServersNegativeTestJSON-584888524 tempest-ListServersNegativeTestJSON-584888524-project-member] Lock "5adbe047-7055-428f-b72b-67023c68cc04" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.502s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.249150] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.886s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.249831] env[62521]: ERROR nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b5049c47-da79-4cb5-979e-f6d73ac5c55a, please check neutron logs for more information. [ 757.249831] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] Traceback (most recent call last): [ 757.249831] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.249831] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] self.driver.spawn(context, instance, image_meta, [ 757.249831] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 757.249831] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.249831] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.249831] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] vm_ref = self.build_virtual_machine(instance, [ 757.249831] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.249831] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.249831] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] for vif in network_info: [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] return self._sync_wrapper(fn, *args, **kwargs) [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] self.wait() [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] self[:] = self._gt.wait() [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] return self._exit_event.wait() [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] result = hub.switch() [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 757.250221] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] return self.greenlet.switch() [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] result = function(*args, **kwargs) [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] return func(*args, **kwargs) [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] raise e [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] nwinfo = self.network_api.allocate_for_instance( [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] created_port_ids = self._update_ports_for_instance( [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] with excutils.save_and_reraise_exception(): [ 757.250581] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] self.force_reraise() [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] raise self.value [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] updated_port = self._update_port( [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] _ensure_no_port_binding_failure(port) [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] raise exception.PortBindingFailed(port_id=port['id']) [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] nova.exception.PortBindingFailed: Binding failed for port b5049c47-da79-4cb5-979e-f6d73ac5c55a, please check neutron logs for more information. [ 757.250934] env[62521]: ERROR nova.compute.manager [instance: 86727bdc-342a-429c-85ce-adec8823b799] [ 757.251247] env[62521]: DEBUG nova.compute.utils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Binding failed for port b5049c47-da79-4cb5-979e-f6d73ac5c55a, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 757.251755] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.235s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.253324] env[62521]: INFO nova.compute.claims [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.257432] env[62521]: DEBUG nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Build of instance 86727bdc-342a-429c-85ce-adec8823b799 was re-scheduled: Binding failed for port b5049c47-da79-4cb5-979e-f6d73ac5c55a, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 757.257569] env[62521]: DEBUG nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 757.257718] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquiring lock "refresh_cache-86727bdc-342a-429c-85ce-adec8823b799" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.257868] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Acquired lock "refresh_cache-86727bdc-342a-429c-85ce-adec8823b799" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.258036] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.520629] env[62521]: INFO nova.compute.manager [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] [instance: 72e3b843-c54a-4de7-ace9-ee25c5df14bd] Took 1.02 seconds to deallocate network for instance. [ 757.665223] env[62521]: DEBUG nova.compute.manager [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.774033] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.866469] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.188456] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.371066] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Releasing lock "refresh_cache-86727bdc-342a-429c-85ce-adec8823b799" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.371405] env[62521]: DEBUG nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 758.371690] env[62521]: DEBUG nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.372747] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.387932] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.553421] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c121102b-5b6a-44b1-adf0-c0113e54194a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.559130] env[62521]: INFO nova.scheduler.client.report [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Deleted allocations for instance 72e3b843-c54a-4de7-ace9-ee25c5df14bd [ 758.568250] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6ff3a7-c1fe-4210-b7a1-afbc46d131f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.598882] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc4a33f-0248-4956-8145-333f3e2b3650 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.608193] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25ff95b-bee3-4afd-a215-168590df0c27 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.622897] env[62521]: DEBUG nova.compute.provider_tree [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.892819] env[62521]: DEBUG nova.network.neutron [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.072090] env[62521]: DEBUG oslo_concurrency.lockutils [None req-de850da4-e98a-408d-a88c-16ce33465931 tempest-ListServerFiltersTestJSON-597941619 tempest-ListServerFiltersTestJSON-597941619-project-member] Lock "72e3b843-c54a-4de7-ace9-ee25c5df14bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.758s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.126006] env[62521]: DEBUG nova.scheduler.client.report [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.199416] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.199649] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.395174] env[62521]: INFO nova.compute.manager [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] [instance: 86727bdc-342a-429c-85ce-adec8823b799] Took 1.02 seconds to deallocate network for instance. [ 759.575343] env[62521]: DEBUG nova.compute.manager [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.631314] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.631810] env[62521]: DEBUG nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 759.635505] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.741s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.637122] env[62521]: INFO nova.compute.claims [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.704143] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.704237] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 759.705251] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Rebuilding the list of instances to heal {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 760.099440] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.137708] env[62521]: DEBUG nova.compute.utils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.139503] env[62521]: DEBUG nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 760.139692] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 760.207312] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 760.207554] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 760.207690] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 760.207816] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Didn't find any instances for network info cache update. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 760.208023] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.208183] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.208328] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.208472] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.208605] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.208743] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.208866] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 760.209015] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.362357] env[62521]: DEBUG nova.policy [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4cdc04a54564499892146525ba8b1f7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d80f916ee8d4b039d1ca4a1faade851', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 760.436225] env[62521]: INFO nova.scheduler.client.report [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Deleted allocations for instance 86727bdc-342a-429c-85ce-adec8823b799 [ 760.640404] env[62521]: DEBUG nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.713783] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.946339] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29c2ddca-7f84-435d-9c8c-3046a78c5b78 tempest-SecurityGroupsTestJSON-1413288972 tempest-SecurityGroupsTestJSON-1413288972-project-member] Lock "86727bdc-342a-429c-85ce-adec8823b799" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.122s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.956898] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Successfully created port: 16a811d0-87b7-45a0-b083-cb9a515738c4 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.987939] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e58bd3f-3ebd-48e8-9355-516a58863a90 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.996987] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1ca941-b7ef-44b5-97cf-63dd1dad3053 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.034501] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f08c90e-c43d-4425-a7c5-fc1e6caf6396 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.043450] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48bd719c-2b31-40c2-88cb-4fad23fc1aa5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.056835] env[62521]: DEBUG nova.compute.provider_tree [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.450598] env[62521]: DEBUG nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 761.559978] env[62521]: DEBUG nova.scheduler.client.report [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.655478] env[62521]: DEBUG nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 761.683380] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.683642] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.683798] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.683981] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.684139] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.684287] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.684494] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.684649] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.684814] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.684973] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.685162] env[62521]: DEBUG nova.virt.hardware [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.686055] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b88f4e-f7fa-469d-974f-ed32f5a4aec9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.693828] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e57190c-0ee4-4385-88f2-46240c09e43d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.974467] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.065788] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.066981] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.952s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.268444] env[62521]: DEBUG nova.compute.manager [req-513983ce-b766-402e-9c80-49f0fc4e7979 req-86c7cb8b-e69c-4355-9a42-f71df4ebd17a service nova] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Received event network-changed-16a811d0-87b7-45a0-b083-cb9a515738c4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.268713] env[62521]: DEBUG nova.compute.manager [req-513983ce-b766-402e-9c80-49f0fc4e7979 req-86c7cb8b-e69c-4355-9a42-f71df4ebd17a service nova] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Refreshing instance network info cache due to event network-changed-16a811d0-87b7-45a0-b083-cb9a515738c4. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 762.268886] env[62521]: DEBUG oslo_concurrency.lockutils [req-513983ce-b766-402e-9c80-49f0fc4e7979 req-86c7cb8b-e69c-4355-9a42-f71df4ebd17a service nova] Acquiring lock "refresh_cache-0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.268987] env[62521]: DEBUG oslo_concurrency.lockutils [req-513983ce-b766-402e-9c80-49f0fc4e7979 req-86c7cb8b-e69c-4355-9a42-f71df4ebd17a service nova] Acquired lock "refresh_cache-0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.269164] env[62521]: DEBUG nova.network.neutron [req-513983ce-b766-402e-9c80-49f0fc4e7979 req-86c7cb8b-e69c-4355-9a42-f71df4ebd17a service nova] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Refreshing network info cache for port 16a811d0-87b7-45a0-b083-cb9a515738c4 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 762.520730] env[62521]: ERROR nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 16a811d0-87b7-45a0-b083-cb9a515738c4, please check neutron logs for more information. [ 762.520730] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 762.520730] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.520730] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 762.520730] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.520730] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 762.520730] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.520730] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 762.520730] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.520730] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 762.520730] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.520730] env[62521]: ERROR nova.compute.manager raise self.value [ 762.520730] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.520730] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 762.520730] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.520730] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 762.521547] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.521547] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 762.521547] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 16a811d0-87b7-45a0-b083-cb9a515738c4, please check neutron logs for more information. [ 762.521547] env[62521]: ERROR nova.compute.manager [ 762.521547] env[62521]: Traceback (most recent call last): [ 762.521547] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 762.521547] env[62521]: listener.cb(fileno) [ 762.521547] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.521547] env[62521]: result = function(*args, **kwargs) [ 762.521547] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.521547] env[62521]: return func(*args, **kwargs) [ 762.521547] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.521547] env[62521]: raise e [ 762.521547] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.521547] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 762.521547] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.521547] env[62521]: created_port_ids = self._update_ports_for_instance( [ 762.521547] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.521547] env[62521]: with excutils.save_and_reraise_exception(): [ 762.521547] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.521547] env[62521]: self.force_reraise() [ 762.521547] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.521547] env[62521]: raise self.value [ 762.521547] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.521547] env[62521]: updated_port = self._update_port( [ 762.521547] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.521547] env[62521]: _ensure_no_port_binding_failure(port) [ 762.521547] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.521547] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 762.522597] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 16a811d0-87b7-45a0-b083-cb9a515738c4, please check neutron logs for more information. [ 762.522597] env[62521]: Removing descriptor: 15 [ 762.522597] env[62521]: ERROR nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 16a811d0-87b7-45a0-b083-cb9a515738c4, please check neutron logs for more information. [ 762.522597] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Traceback (most recent call last): [ 762.522597] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 762.522597] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] yield resources [ 762.522597] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.522597] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] self.driver.spawn(context, instance, image_meta, [ 762.522597] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 762.522597] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.522597] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.522597] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] vm_ref = self.build_virtual_machine(instance, [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] for vif in network_info: [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] return self._sync_wrapper(fn, *args, **kwargs) [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] self.wait() [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] self[:] = self._gt.wait() [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] return self._exit_event.wait() [ 762.523078] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] result = hub.switch() [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] return self.greenlet.switch() [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] result = function(*args, **kwargs) [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] return func(*args, **kwargs) [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] raise e [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] nwinfo = self.network_api.allocate_for_instance( [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.523451] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] created_port_ids = self._update_ports_for_instance( [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] with excutils.save_and_reraise_exception(): [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] self.force_reraise() [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] raise self.value [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] updated_port = self._update_port( [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] _ensure_no_port_binding_failure(port) [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.523882] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] raise exception.PortBindingFailed(port_id=port['id']) [ 762.524231] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] nova.exception.PortBindingFailed: Binding failed for port 16a811d0-87b7-45a0-b083-cb9a515738c4, please check neutron logs for more information. [ 762.524231] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] [ 762.524231] env[62521]: INFO nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Terminating instance [ 762.524850] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Acquiring lock "refresh_cache-0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.572428] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Acquiring lock "2dae04bc-8dd4-4eb8-ad24-f3200331685f" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.572710] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Lock "2dae04bc-8dd4-4eb8-ad24-f3200331685f" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.794548] env[62521]: DEBUG nova.network.neutron [req-513983ce-b766-402e-9c80-49f0fc4e7979 req-86c7cb8b-e69c-4355-9a42-f71df4ebd17a service nova] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.819296] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3348c1-5a1f-462c-b551-b49209badfe1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.828175] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2390d28-452d-45c4-b9cf-96e23f7316d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.858910] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524322f2-50ba-4713-b91e-1c2bad9dfc7b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.866343] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984476d8-2c34-42f7-8362-ed9a3b4933df {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.880595] env[62521]: DEBUG nova.compute.provider_tree [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.948841] env[62521]: DEBUG nova.network.neutron [req-513983ce-b766-402e-9c80-49f0fc4e7979 req-86c7cb8b-e69c-4355-9a42-f71df4ebd17a service nova] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.076128] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Lock "2dae04bc-8dd4-4eb8-ad24-f3200331685f" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.503s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.078350] env[62521]: DEBUG nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 763.385509] env[62521]: DEBUG nova.scheduler.client.report [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.451963] env[62521]: DEBUG oslo_concurrency.lockutils [req-513983ce-b766-402e-9c80-49f0fc4e7979 req-86c7cb8b-e69c-4355-9a42-f71df4ebd17a service nova] Releasing lock "refresh_cache-0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.452327] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Acquired lock "refresh_cache-0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.452550] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.581990] env[62521]: DEBUG nova.compute.utils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.583397] env[62521]: DEBUG nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 763.583567] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 763.693925] env[62521]: DEBUG nova.policy [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ae45536cb624e45811d1aa06e2e1c58', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd75697fd8970468fb135fa55396411e8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.890406] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.823s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.891083] env[62521]: ERROR nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d72099b7-8aec-49f1-b2c8-c1b080e3f10d, please check neutron logs for more information. [ 763.891083] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Traceback (most recent call last): [ 763.891083] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 763.891083] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] self.driver.spawn(context, instance, image_meta, [ 763.891083] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 763.891083] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.891083] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.891083] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] vm_ref = self.build_virtual_machine(instance, [ 763.891083] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.891083] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.891083] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] for vif in network_info: [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] return self._sync_wrapper(fn, *args, **kwargs) [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] self.wait() [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] self[:] = self._gt.wait() [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] return self._exit_event.wait() [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] result = hub.switch() [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 763.891797] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] return self.greenlet.switch() [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] result = function(*args, **kwargs) [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] return func(*args, **kwargs) [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] raise e [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] nwinfo = self.network_api.allocate_for_instance( [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] created_port_ids = self._update_ports_for_instance( [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] with excutils.save_and_reraise_exception(): [ 763.893093] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] self.force_reraise() [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] raise self.value [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] updated_port = self._update_port( [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] _ensure_no_port_binding_failure(port) [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] raise exception.PortBindingFailed(port_id=port['id']) [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] nova.exception.PortBindingFailed: Binding failed for port d72099b7-8aec-49f1-b2c8-c1b080e3f10d, please check neutron logs for more information. [ 763.893871] env[62521]: ERROR nova.compute.manager [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] [ 763.894387] env[62521]: DEBUG nova.compute.utils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Binding failed for port d72099b7-8aec-49f1-b2c8-c1b080e3f10d, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 763.894387] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.480s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.894500] env[62521]: INFO nova.compute.claims [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 763.897422] env[62521]: DEBUG nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Build of instance 277ecbd7-e770-4ea0-ae2e-b49598428d0f was re-scheduled: Binding failed for port d72099b7-8aec-49f1-b2c8-c1b080e3f10d, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 763.897867] env[62521]: DEBUG nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 763.901308] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Acquiring lock "refresh_cache-277ecbd7-e770-4ea0-ae2e-b49598428d0f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.902064] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Acquired lock "refresh_cache-277ecbd7-e770-4ea0-ae2e-b49598428d0f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.902064] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.978986] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.061811] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.090342] env[62521]: DEBUG nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 764.296304] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Successfully created port: 4365d026-b662-4559-b65f-4d64ad164f6e {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.404504] env[62521]: DEBUG nova.compute.manager [req-db3b49ff-7708-4b60-8592-4f23ac7377ee req-bf609489-e5e5-4d42-9cc6-e4f3a3cddf00 service nova] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Received event network-vif-deleted-16a811d0-87b7-45a0-b083-cb9a515738c4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.435580] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.502468] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.564645] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Releasing lock "refresh_cache-0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.565805] env[62521]: DEBUG nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 764.566016] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 764.566419] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0e3a517c-0dc9-49ba-9c38-8ad015233733 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.575436] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca694461-4090-4db1-af18-60f515aa4c08 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.600453] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a could not be found. [ 764.600722] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 764.600949] env[62521]: INFO nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 764.601233] env[62521]: DEBUG oslo.service.loopingcall [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.601679] env[62521]: DEBUG nova.compute.manager [-] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.601812] env[62521]: DEBUG nova.network.neutron [-] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.617119] env[62521]: DEBUG nova.network.neutron [-] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.005553] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Releasing lock "refresh_cache-277ecbd7-e770-4ea0-ae2e-b49598428d0f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.005776] env[62521]: DEBUG nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 765.005867] env[62521]: DEBUG nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.005962] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.025961] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.102928] env[62521]: DEBUG nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 765.121111] env[62521]: DEBUG nova.network.neutron [-] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.131543] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 765.131932] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 765.132207] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.132513] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 765.132718] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.132876] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 765.133101] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 765.133269] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 765.133439] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 765.133600] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 765.133813] env[62521]: DEBUG nova.virt.hardware [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 765.134685] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cecfd3bd-2aa5-430f-aacb-01a40d015110 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.143770] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a68938-f357-4b8f-98c2-1a8e3d290a0c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.218618] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c49cf0-800c-48c5-9ef5-c09872434734 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.227091] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd648fd9-7033-4593-8990-612e922f666a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.264066] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2fb8428-bcd2-4ff2-8fba-5c5e0fb54c01 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.274932] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a7b840-99fa-4de1-b5fb-65e604abdc41 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.288027] env[62521]: DEBUG nova.compute.provider_tree [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.529034] env[62521]: DEBUG nova.network.neutron [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.602833] env[62521]: ERROR nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4365d026-b662-4559-b65f-4d64ad164f6e, please check neutron logs for more information. [ 765.602833] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 765.602833] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.602833] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 765.602833] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.602833] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 765.602833] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.602833] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 765.602833] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.602833] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 765.602833] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.602833] env[62521]: ERROR nova.compute.manager raise self.value [ 765.602833] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.602833] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 765.602833] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.602833] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 765.603404] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.603404] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 765.603404] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4365d026-b662-4559-b65f-4d64ad164f6e, please check neutron logs for more information. [ 765.603404] env[62521]: ERROR nova.compute.manager [ 765.603404] env[62521]: Traceback (most recent call last): [ 765.603404] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 765.603404] env[62521]: listener.cb(fileno) [ 765.603404] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.603404] env[62521]: result = function(*args, **kwargs) [ 765.603404] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 765.603404] env[62521]: return func(*args, **kwargs) [ 765.603404] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.603404] env[62521]: raise e [ 765.603404] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.603404] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 765.603404] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.603404] env[62521]: created_port_ids = self._update_ports_for_instance( [ 765.603404] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.603404] env[62521]: with excutils.save_and_reraise_exception(): [ 765.603404] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.603404] env[62521]: self.force_reraise() [ 765.603404] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.603404] env[62521]: raise self.value [ 765.603404] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.603404] env[62521]: updated_port = self._update_port( [ 765.603404] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.603404] env[62521]: _ensure_no_port_binding_failure(port) [ 765.603404] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.603404] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 765.604326] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 4365d026-b662-4559-b65f-4d64ad164f6e, please check neutron logs for more information. [ 765.604326] env[62521]: Removing descriptor: 15 [ 765.604326] env[62521]: ERROR nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4365d026-b662-4559-b65f-4d64ad164f6e, please check neutron logs for more information. [ 765.604326] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Traceback (most recent call last): [ 765.604326] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 765.604326] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] yield resources [ 765.604326] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.604326] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] self.driver.spawn(context, instance, image_meta, [ 765.604326] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 765.604326] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.604326] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.604326] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] vm_ref = self.build_virtual_machine(instance, [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] for vif in network_info: [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] return self._sync_wrapper(fn, *args, **kwargs) [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] self.wait() [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] self[:] = self._gt.wait() [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] return self._exit_event.wait() [ 765.604727] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] result = hub.switch() [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] return self.greenlet.switch() [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] result = function(*args, **kwargs) [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] return func(*args, **kwargs) [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] raise e [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] nwinfo = self.network_api.allocate_for_instance( [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.605158] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] created_port_ids = self._update_ports_for_instance( [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] with excutils.save_and_reraise_exception(): [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] self.force_reraise() [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] raise self.value [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] updated_port = self._update_port( [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] _ensure_no_port_binding_failure(port) [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.605637] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] raise exception.PortBindingFailed(port_id=port['id']) [ 765.606101] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] nova.exception.PortBindingFailed: Binding failed for port 4365d026-b662-4559-b65f-4d64ad164f6e, please check neutron logs for more information. [ 765.606101] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] [ 765.606101] env[62521]: INFO nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Terminating instance [ 765.606101] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Acquiring lock "refresh_cache-32c0268b-eee6-4e45-b9c3-077ec81bd314" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.606262] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Acquired lock "refresh_cache-32c0268b-eee6-4e45-b9c3-077ec81bd314" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.606398] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.623245] env[62521]: INFO nova.compute.manager [-] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Took 1.02 seconds to deallocate network for instance. [ 765.625123] env[62521]: DEBUG nova.compute.claims [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 765.625322] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.793863] env[62521]: DEBUG nova.scheduler.client.report [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.033893] env[62521]: INFO nova.compute.manager [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] [instance: 277ecbd7-e770-4ea0-ae2e-b49598428d0f] Took 1.03 seconds to deallocate network for instance. [ 766.130161] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.257872] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.298908] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.299438] env[62521]: DEBUG nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.301905] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.972s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.303387] env[62521]: INFO nova.compute.claims [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.430849] env[62521]: DEBUG nova.compute.manager [req-b1253cdf-7246-47c3-a495-33217dba2fbd req-5568bead-90cd-4c1d-8df2-75118c4a94db service nova] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Received event network-changed-4365d026-b662-4559-b65f-4d64ad164f6e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.431055] env[62521]: DEBUG nova.compute.manager [req-b1253cdf-7246-47c3-a495-33217dba2fbd req-5568bead-90cd-4c1d-8df2-75118c4a94db service nova] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Refreshing instance network info cache due to event network-changed-4365d026-b662-4559-b65f-4d64ad164f6e. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 766.431247] env[62521]: DEBUG oslo_concurrency.lockutils [req-b1253cdf-7246-47c3-a495-33217dba2fbd req-5568bead-90cd-4c1d-8df2-75118c4a94db service nova] Acquiring lock "refresh_cache-32c0268b-eee6-4e45-b9c3-077ec81bd314" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.760552] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Releasing lock "refresh_cache-32c0268b-eee6-4e45-b9c3-077ec81bd314" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.760975] env[62521]: DEBUG nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 766.761179] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 766.761480] env[62521]: DEBUG oslo_concurrency.lockutils [req-b1253cdf-7246-47c3-a495-33217dba2fbd req-5568bead-90cd-4c1d-8df2-75118c4a94db service nova] Acquired lock "refresh_cache-32c0268b-eee6-4e45-b9c3-077ec81bd314" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.761679] env[62521]: DEBUG nova.network.neutron [req-b1253cdf-7246-47c3-a495-33217dba2fbd req-5568bead-90cd-4c1d-8df2-75118c4a94db service nova] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Refreshing network info cache for port 4365d026-b662-4559-b65f-4d64ad164f6e {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 766.762701] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fbdf1af6-a039-4219-a522-cb9b8b71c67e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.771870] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941f6a3f-7167-45b5-8a04-a197609bcda4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.793757] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 32c0268b-eee6-4e45-b9c3-077ec81bd314 could not be found. [ 766.793899] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 766.793964] env[62521]: INFO nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Took 0.03 seconds to destroy the instance on the hypervisor. [ 766.794217] env[62521]: DEBUG oslo.service.loopingcall [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.794446] env[62521]: DEBUG nova.compute.manager [-] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.794538] env[62521]: DEBUG nova.network.neutron [-] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.807772] env[62521]: DEBUG nova.compute.utils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 766.811590] env[62521]: DEBUG nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 766.811761] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 766.818275] env[62521]: DEBUG nova.network.neutron [-] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.858052] env[62521]: DEBUG nova.policy [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4389087709a465da302f80143708f13', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2fe41e12d5df4750b8169a148e81aec3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.084771] env[62521]: INFO nova.scheduler.client.report [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Deleted allocations for instance 277ecbd7-e770-4ea0-ae2e-b49598428d0f [ 767.148815] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Successfully created port: 4bee5029-6582-4a51-9759-3d91534d13b5 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.304557] env[62521]: DEBUG nova.network.neutron [req-b1253cdf-7246-47c3-a495-33217dba2fbd req-5568bead-90cd-4c1d-8df2-75118c4a94db service nova] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.312675] env[62521]: DEBUG nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.324432] env[62521]: DEBUG nova.network.neutron [-] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.536702] env[62521]: DEBUG nova.network.neutron [req-b1253cdf-7246-47c3-a495-33217dba2fbd req-5568bead-90cd-4c1d-8df2-75118c4a94db service nova] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.595623] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67b9c51c-d326-4926-8207-0b68c6064969 tempest-ServersAdminNegativeTestJSON-1018915910 tempest-ServersAdminNegativeTestJSON-1018915910-project-member] Lock "277ecbd7-e770-4ea0-ae2e-b49598428d0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.215s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.671760] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf815118-9284-41f0-b185-a1e60087186e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.679996] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c76022d-e53a-4a6c-ac4c-ae99b6e03509 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.715411] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80730a1e-f410-4d7a-9163-cf4f0f44e80b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.723605] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4197140-b0b5-4e7d-95bf-9fb9926a80b3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.738853] env[62521]: DEBUG nova.compute.provider_tree [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.831384] env[62521]: INFO nova.compute.manager [-] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Took 1.04 seconds to deallocate network for instance. [ 767.836318] env[62521]: DEBUG nova.compute.claims [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 767.836369] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.040748] env[62521]: DEBUG oslo_concurrency.lockutils [req-b1253cdf-7246-47c3-a495-33217dba2fbd req-5568bead-90cd-4c1d-8df2-75118c4a94db service nova] Releasing lock "refresh_cache-32c0268b-eee6-4e45-b9c3-077ec81bd314" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.040748] env[62521]: DEBUG nova.compute.manager [req-b1253cdf-7246-47c3-a495-33217dba2fbd req-5568bead-90cd-4c1d-8df2-75118c4a94db service nova] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Received event network-vif-deleted-4365d026-b662-4559-b65f-4d64ad164f6e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.084462] env[62521]: DEBUG nova.compute.manager [req-95109ee0-2b24-4b7f-9e27-47490e137b48 req-cb598925-bdb1-4a66-9293-f0adbbb3e8e9 service nova] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Received event network-changed-4bee5029-6582-4a51-9759-3d91534d13b5 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.084663] env[62521]: DEBUG nova.compute.manager [req-95109ee0-2b24-4b7f-9e27-47490e137b48 req-cb598925-bdb1-4a66-9293-f0adbbb3e8e9 service nova] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Refreshing instance network info cache due to event network-changed-4bee5029-6582-4a51-9759-3d91534d13b5. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 768.084879] env[62521]: DEBUG oslo_concurrency.lockutils [req-95109ee0-2b24-4b7f-9e27-47490e137b48 req-cb598925-bdb1-4a66-9293-f0adbbb3e8e9 service nova] Acquiring lock "refresh_cache-63abbe0a-af36-4679-8cf1-87f9e9abf7b9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.085121] env[62521]: DEBUG oslo_concurrency.lockutils [req-95109ee0-2b24-4b7f-9e27-47490e137b48 req-cb598925-bdb1-4a66-9293-f0adbbb3e8e9 service nova] Acquired lock "refresh_cache-63abbe0a-af36-4679-8cf1-87f9e9abf7b9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.085179] env[62521]: DEBUG nova.network.neutron [req-95109ee0-2b24-4b7f-9e27-47490e137b48 req-cb598925-bdb1-4a66-9293-f0adbbb3e8e9 service nova] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Refreshing network info cache for port 4bee5029-6582-4a51-9759-3d91534d13b5 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 768.098518] env[62521]: DEBUG nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 768.242370] env[62521]: DEBUG nova.scheduler.client.report [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.251370] env[62521]: ERROR nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4bee5029-6582-4a51-9759-3d91534d13b5, please check neutron logs for more information. [ 768.251370] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 768.251370] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.251370] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 768.251370] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.251370] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 768.251370] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.251370] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 768.251370] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.251370] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 768.251370] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.251370] env[62521]: ERROR nova.compute.manager raise self.value [ 768.251370] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.251370] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 768.251370] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.251370] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 768.252325] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.252325] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 768.252325] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4bee5029-6582-4a51-9759-3d91534d13b5, please check neutron logs for more information. [ 768.252325] env[62521]: ERROR nova.compute.manager [ 768.252325] env[62521]: Traceback (most recent call last): [ 768.252325] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 768.252325] env[62521]: listener.cb(fileno) [ 768.252325] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.252325] env[62521]: result = function(*args, **kwargs) [ 768.252325] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 768.252325] env[62521]: return func(*args, **kwargs) [ 768.252325] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 768.252325] env[62521]: raise e [ 768.252325] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.252325] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 768.252325] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.252325] env[62521]: created_port_ids = self._update_ports_for_instance( [ 768.252325] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.252325] env[62521]: with excutils.save_and_reraise_exception(): [ 768.252325] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.252325] env[62521]: self.force_reraise() [ 768.252325] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.252325] env[62521]: raise self.value [ 768.252325] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.252325] env[62521]: updated_port = self._update_port( [ 768.252325] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.252325] env[62521]: _ensure_no_port_binding_failure(port) [ 768.252325] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.252325] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 768.253193] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 4bee5029-6582-4a51-9759-3d91534d13b5, please check neutron logs for more information. [ 768.253193] env[62521]: Removing descriptor: 18 [ 768.330743] env[62521]: DEBUG nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 768.364515] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.364958] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.364958] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.365097] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.365328] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.365510] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.365740] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.365852] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.366028] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.366198] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.366400] env[62521]: DEBUG nova.virt.hardware [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.368693] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a9615a-251b-432e-87b8-2f235999312e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.375453] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6cc079d-0a32-45f6-a715-ab72c1a0cb40 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.390710] env[62521]: ERROR nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4bee5029-6582-4a51-9759-3d91534d13b5, please check neutron logs for more information. [ 768.390710] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Traceback (most recent call last): [ 768.390710] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 768.390710] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] yield resources [ 768.390710] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 768.390710] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] self.driver.spawn(context, instance, image_meta, [ 768.390710] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 768.390710] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 768.390710] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 768.390710] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] vm_ref = self.build_virtual_machine(instance, [ 768.390710] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] for vif in network_info: [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] return self._sync_wrapper(fn, *args, **kwargs) [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] self.wait() [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] self[:] = self._gt.wait() [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] return self._exit_event.wait() [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 768.391124] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] current.throw(*self._exc) [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] result = function(*args, **kwargs) [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] return func(*args, **kwargs) [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] raise e [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] nwinfo = self.network_api.allocate_for_instance( [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] created_port_ids = self._update_ports_for_instance( [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] with excutils.save_and_reraise_exception(): [ 768.391551] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] self.force_reraise() [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] raise self.value [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] updated_port = self._update_port( [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] _ensure_no_port_binding_failure(port) [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] raise exception.PortBindingFailed(port_id=port['id']) [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] nova.exception.PortBindingFailed: Binding failed for port 4bee5029-6582-4a51-9759-3d91534d13b5, please check neutron logs for more information. [ 768.392011] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] [ 768.392011] env[62521]: INFO nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Terminating instance [ 768.393162] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Acquiring lock "refresh_cache-63abbe0a-af36-4679-8cf1-87f9e9abf7b9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.606900] env[62521]: DEBUG nova.network.neutron [req-95109ee0-2b24-4b7f-9e27-47490e137b48 req-cb598925-bdb1-4a66-9293-f0adbbb3e8e9 service nova] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.624624] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.729586] env[62521]: DEBUG nova.network.neutron [req-95109ee0-2b24-4b7f-9e27-47490e137b48 req-cb598925-bdb1-4a66-9293-f0adbbb3e8e9 service nova] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.753852] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.754391] env[62521]: DEBUG nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.761081] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.414s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.762359] env[62521]: INFO nova.compute.claims [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.232686] env[62521]: DEBUG oslo_concurrency.lockutils [req-95109ee0-2b24-4b7f-9e27-47490e137b48 req-cb598925-bdb1-4a66-9293-f0adbbb3e8e9 service nova] Releasing lock "refresh_cache-63abbe0a-af36-4679-8cf1-87f9e9abf7b9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.233337] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Acquired lock "refresh_cache-63abbe0a-af36-4679-8cf1-87f9e9abf7b9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.233874] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.258981] env[62521]: DEBUG nova.compute.utils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.260343] env[62521]: DEBUG nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 769.260516] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 769.325556] env[62521]: DEBUG nova.policy [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6452a974466a460fb7b2ccaf4e73cc5b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd1ea05250f5497c80a295a98b155b6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.389015] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquiring lock "92629ebf-eb50-4862-abe2-01e210e302ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.389015] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "92629ebf-eb50-4862-abe2-01e210e302ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.675203] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Successfully created port: eaa99398-192b-4f83-a3fb-a0ac0164fa23 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.756998] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.763719] env[62521]: DEBUG nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.975069] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.079506] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99da265-eae7-4279-9a8f-90fbbc95731a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.087513] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe5915e-fe8b-4fc6-802a-20a366f2384e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.123022] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dfb7157-50a2-495f-b43c-7f0c84519c82 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.129724] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255482f1-07bf-481f-ba00-ea192891eba0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.145042] env[62521]: DEBUG nova.compute.provider_tree [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.171167] env[62521]: DEBUG nova.compute.manager [req-c3e8e453-ff9b-426e-a160-d80067d5bf6e req-a16fd79e-de1d-40de-b743-43d750a54a4a service nova] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Received event network-vif-deleted-4bee5029-6582-4a51-9759-3d91534d13b5 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.478568] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Releasing lock "refresh_cache-63abbe0a-af36-4679-8cf1-87f9e9abf7b9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.479007] env[62521]: DEBUG nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 770.479218] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 770.479514] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-58e410cb-7bbc-48a2-aeec-022f9e8c8031 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.488550] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567b44b5-2d3b-4ce0-8bb5-8dc1393cfe0c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.510672] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 63abbe0a-af36-4679-8cf1-87f9e9abf7b9 could not be found. [ 770.510883] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 770.511071] env[62521]: INFO nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 770.511327] env[62521]: DEBUG oslo.service.loopingcall [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.511534] env[62521]: DEBUG nova.compute.manager [-] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.511625] env[62521]: DEBUG nova.network.neutron [-] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.534322] env[62521]: DEBUG nova.network.neutron [-] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.558507] env[62521]: ERROR nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eaa99398-192b-4f83-a3fb-a0ac0164fa23, please check neutron logs for more information. [ 770.558507] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 770.558507] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.558507] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 770.558507] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.558507] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 770.558507] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.558507] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 770.558507] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.558507] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 770.558507] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.558507] env[62521]: ERROR nova.compute.manager raise self.value [ 770.558507] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.558507] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 770.558507] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.558507] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 770.558982] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.558982] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 770.558982] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eaa99398-192b-4f83-a3fb-a0ac0164fa23, please check neutron logs for more information. [ 770.558982] env[62521]: ERROR nova.compute.manager [ 770.558982] env[62521]: Traceback (most recent call last): [ 770.558982] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 770.558982] env[62521]: listener.cb(fileno) [ 770.558982] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.558982] env[62521]: result = function(*args, **kwargs) [ 770.558982] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 770.558982] env[62521]: return func(*args, **kwargs) [ 770.558982] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.558982] env[62521]: raise e [ 770.558982] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.558982] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 770.558982] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.558982] env[62521]: created_port_ids = self._update_ports_for_instance( [ 770.558982] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.558982] env[62521]: with excutils.save_and_reraise_exception(): [ 770.558982] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.558982] env[62521]: self.force_reraise() [ 770.558982] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.558982] env[62521]: raise self.value [ 770.558982] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.558982] env[62521]: updated_port = self._update_port( [ 770.558982] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.558982] env[62521]: _ensure_no_port_binding_failure(port) [ 770.558982] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.558982] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 770.559948] env[62521]: nova.exception.PortBindingFailed: Binding failed for port eaa99398-192b-4f83-a3fb-a0ac0164fa23, please check neutron logs for more information. [ 770.559948] env[62521]: Removing descriptor: 18 [ 770.648028] env[62521]: DEBUG nova.scheduler.client.report [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.773712] env[62521]: DEBUG nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.802170] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.802376] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.802533] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.802714] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.802859] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.803013] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.803383] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.803569] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.803744] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.803906] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.804090] env[62521]: DEBUG nova.virt.hardware [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.805378] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151055a7-b935-4e43-8fdd-918fe3c7bc54 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.813694] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43dc8885-bddc-45a8-b012-0a735b240b8e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.831607] env[62521]: ERROR nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eaa99398-192b-4f83-a3fb-a0ac0164fa23, please check neutron logs for more information. [ 770.831607] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Traceback (most recent call last): [ 770.831607] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 770.831607] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] yield resources [ 770.831607] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 770.831607] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] self.driver.spawn(context, instance, image_meta, [ 770.831607] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 770.831607] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.831607] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.831607] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] vm_ref = self.build_virtual_machine(instance, [ 770.831607] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] for vif in network_info: [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] return self._sync_wrapper(fn, *args, **kwargs) [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] self.wait() [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] self[:] = self._gt.wait() [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] return self._exit_event.wait() [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 770.834763] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] current.throw(*self._exc) [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] result = function(*args, **kwargs) [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] return func(*args, **kwargs) [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] raise e [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] nwinfo = self.network_api.allocate_for_instance( [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] created_port_ids = self._update_ports_for_instance( [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] with excutils.save_and_reraise_exception(): [ 770.835148] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] self.force_reraise() [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] raise self.value [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] updated_port = self._update_port( [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] _ensure_no_port_binding_failure(port) [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] raise exception.PortBindingFailed(port_id=port['id']) [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] nova.exception.PortBindingFailed: Binding failed for port eaa99398-192b-4f83-a3fb-a0ac0164fa23, please check neutron logs for more information. [ 770.835557] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] [ 770.835557] env[62521]: INFO nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Terminating instance [ 770.835968] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Acquiring lock "refresh_cache-f6a27e36-1728-4056-99a6-ed2114df64ce" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.835968] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Acquired lock "refresh_cache-f6a27e36-1728-4056-99a6-ed2114df64ce" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.835968] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.040674] env[62521]: DEBUG nova.network.neutron [-] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.155970] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.156610] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.161008] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.968s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.162415] env[62521]: INFO nova.compute.claims [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.354466] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.429052] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.543543] env[62521]: INFO nova.compute.manager [-] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Took 1.03 seconds to deallocate network for instance. [ 771.545614] env[62521]: DEBUG nova.compute.claims [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 771.545794] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.672018] env[62521]: DEBUG nova.compute.utils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.675588] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 771.675893] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 771.726738] env[62521]: DEBUG nova.policy [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05e0e43e72b74039b813d71aebd21872', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63ec3a6586ee49059b51ccfcaaeab7b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 771.934899] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Releasing lock "refresh_cache-f6a27e36-1728-4056-99a6-ed2114df64ce" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.935385] env[62521]: DEBUG nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.935593] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.935902] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8310d6a8-28b4-4104-a348-267eaddde4dd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.945528] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199d720d-9c99-4686-b6a5-fe469a7d854d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.973875] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f6a27e36-1728-4056-99a6-ed2114df64ce could not be found. [ 771.974127] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.974340] env[62521]: INFO nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Took 0.04 seconds to destroy the instance on the hypervisor. [ 771.974618] env[62521]: DEBUG oslo.service.loopingcall [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.974848] env[62521]: DEBUG nova.compute.manager [-] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.974942] env[62521]: DEBUG nova.network.neutron [-] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.996754] env[62521]: DEBUG nova.network.neutron [-] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.084480] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Successfully created port: c0bc58a3-690e-4f33-8b5a-2c55766167f5 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.177247] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 772.207899] env[62521]: DEBUG nova.compute.manager [req-a7b507e1-d515-49d1-96e0-2f2470ce41b0 req-f8923602-e579-45ca-8072-5705f34aaec7 service nova] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Received event network-changed-eaa99398-192b-4f83-a3fb-a0ac0164fa23 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.208121] env[62521]: DEBUG nova.compute.manager [req-a7b507e1-d515-49d1-96e0-2f2470ce41b0 req-f8923602-e579-45ca-8072-5705f34aaec7 service nova] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Refreshing instance network info cache due to event network-changed-eaa99398-192b-4f83-a3fb-a0ac0164fa23. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 772.208341] env[62521]: DEBUG oslo_concurrency.lockutils [req-a7b507e1-d515-49d1-96e0-2f2470ce41b0 req-f8923602-e579-45ca-8072-5705f34aaec7 service nova] Acquiring lock "refresh_cache-f6a27e36-1728-4056-99a6-ed2114df64ce" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.208487] env[62521]: DEBUG oslo_concurrency.lockutils [req-a7b507e1-d515-49d1-96e0-2f2470ce41b0 req-f8923602-e579-45ca-8072-5705f34aaec7 service nova] Acquired lock "refresh_cache-f6a27e36-1728-4056-99a6-ed2114df64ce" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.208645] env[62521]: DEBUG nova.network.neutron [req-a7b507e1-d515-49d1-96e0-2f2470ce41b0 req-f8923602-e579-45ca-8072-5705f34aaec7 service nova] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Refreshing network info cache for port eaa99398-192b-4f83-a3fb-a0ac0164fa23 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.499799] env[62521]: DEBUG nova.network.neutron [-] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.529621] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8dadcc4-002f-451b-a5df-689c15226215 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.539926] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e395cff5-bf91-48d5-8ffa-8488dfd30f96 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.572864] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676dbdf9-aa2f-41d5-b013-2a95ebab77cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.584218] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c758433b-622a-48ff-a4d1-52b828e72226 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.598107] env[62521]: DEBUG nova.compute.provider_tree [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.740790] env[62521]: DEBUG nova.network.neutron [req-a7b507e1-d515-49d1-96e0-2f2470ce41b0 req-f8923602-e579-45ca-8072-5705f34aaec7 service nova] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.835537] env[62521]: DEBUG nova.network.neutron [req-a7b507e1-d515-49d1-96e0-2f2470ce41b0 req-f8923602-e579-45ca-8072-5705f34aaec7 service nova] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.002638] env[62521]: INFO nova.compute.manager [-] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Took 1.03 seconds to deallocate network for instance. [ 773.005069] env[62521]: DEBUG nova.compute.claims [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 773.005296] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.100581] env[62521]: DEBUG nova.scheduler.client.report [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.186295] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 773.224230] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 773.224480] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 773.224630] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 773.224802] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 773.224941] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 773.225162] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 773.225358] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 773.225583] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 773.225758] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 773.225858] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 773.226029] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 773.226931] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90299ad9-804c-48df-aabf-386c25822da5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.234603] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21ef892-64cf-4f57-a08e-78bc0f3bb062 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.338787] env[62521]: DEBUG oslo_concurrency.lockutils [req-a7b507e1-d515-49d1-96e0-2f2470ce41b0 req-f8923602-e579-45ca-8072-5705f34aaec7 service nova] Releasing lock "refresh_cache-f6a27e36-1728-4056-99a6-ed2114df64ce" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.339040] env[62521]: DEBUG nova.compute.manager [req-a7b507e1-d515-49d1-96e0-2f2470ce41b0 req-f8923602-e579-45ca-8072-5705f34aaec7 service nova] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Received event network-vif-deleted-eaa99398-192b-4f83-a3fb-a0ac0164fa23 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.609636] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.610270] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 773.613346] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.425s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.618019] env[62521]: INFO nova.compute.claims [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.770630] env[62521]: ERROR nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c0bc58a3-690e-4f33-8b5a-2c55766167f5, please check neutron logs for more information. [ 773.770630] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 773.770630] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.770630] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 773.770630] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.770630] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 773.770630] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.770630] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 773.770630] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.770630] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 773.770630] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.770630] env[62521]: ERROR nova.compute.manager raise self.value [ 773.770630] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.770630] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 773.770630] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.770630] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 773.771640] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.771640] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 773.771640] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c0bc58a3-690e-4f33-8b5a-2c55766167f5, please check neutron logs for more information. [ 773.771640] env[62521]: ERROR nova.compute.manager [ 773.771640] env[62521]: Traceback (most recent call last): [ 773.771640] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 773.771640] env[62521]: listener.cb(fileno) [ 773.771640] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.771640] env[62521]: result = function(*args, **kwargs) [ 773.771640] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.771640] env[62521]: return func(*args, **kwargs) [ 773.771640] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.771640] env[62521]: raise e [ 773.771640] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.771640] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 773.771640] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.771640] env[62521]: created_port_ids = self._update_ports_for_instance( [ 773.771640] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.771640] env[62521]: with excutils.save_and_reraise_exception(): [ 773.771640] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.771640] env[62521]: self.force_reraise() [ 773.771640] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.771640] env[62521]: raise self.value [ 773.771640] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.771640] env[62521]: updated_port = self._update_port( [ 773.771640] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.771640] env[62521]: _ensure_no_port_binding_failure(port) [ 773.771640] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.771640] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 773.778720] env[62521]: nova.exception.PortBindingFailed: Binding failed for port c0bc58a3-690e-4f33-8b5a-2c55766167f5, please check neutron logs for more information. [ 773.778720] env[62521]: Removing descriptor: 18 [ 773.778720] env[62521]: ERROR nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c0bc58a3-690e-4f33-8b5a-2c55766167f5, please check neutron logs for more information. [ 773.778720] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Traceback (most recent call last): [ 773.778720] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 773.778720] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] yield resources [ 773.778720] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.778720] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] self.driver.spawn(context, instance, image_meta, [ 773.778720] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 773.778720] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.778720] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.778720] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] vm_ref = self.build_virtual_machine(instance, [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] for vif in network_info: [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] return self._sync_wrapper(fn, *args, **kwargs) [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] self.wait() [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] self[:] = self._gt.wait() [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] return self._exit_event.wait() [ 773.779197] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] result = hub.switch() [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] return self.greenlet.switch() [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] result = function(*args, **kwargs) [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] return func(*args, **kwargs) [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] raise e [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] nwinfo = self.network_api.allocate_for_instance( [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.779573] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] created_port_ids = self._update_ports_for_instance( [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] with excutils.save_and_reraise_exception(): [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] self.force_reraise() [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] raise self.value [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] updated_port = self._update_port( [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] _ensure_no_port_binding_failure(port) [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.779984] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] raise exception.PortBindingFailed(port_id=port['id']) [ 773.780349] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] nova.exception.PortBindingFailed: Binding failed for port c0bc58a3-690e-4f33-8b5a-2c55766167f5, please check neutron logs for more information. [ 773.780349] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] [ 773.780349] env[62521]: INFO nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Terminating instance [ 773.780349] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "refresh_cache-6be7d826-e19b-431e-b714-25e646e6b7ca" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.780349] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquired lock "refresh_cache-6be7d826-e19b-431e-b714-25e646e6b7ca" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.780349] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 774.125581] env[62521]: DEBUG nova.compute.utils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.127347] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.127347] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 774.199024] env[62521]: DEBUG nova.policy [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05e0e43e72b74039b813d71aebd21872', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63ec3a6586ee49059b51ccfcaaeab7b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 774.234269] env[62521]: DEBUG nova.compute.manager [req-63c6f548-bdb2-4be5-ba1c-a233442c2bf9 req-e588b14a-79a2-48a0-82b9-c5132deddd65 service nova] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Received event network-changed-c0bc58a3-690e-4f33-8b5a-2c55766167f5 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.234471] env[62521]: DEBUG nova.compute.manager [req-63c6f548-bdb2-4be5-ba1c-a233442c2bf9 req-e588b14a-79a2-48a0-82b9-c5132deddd65 service nova] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Refreshing instance network info cache due to event network-changed-c0bc58a3-690e-4f33-8b5a-2c55766167f5. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 774.235102] env[62521]: DEBUG oslo_concurrency.lockutils [req-63c6f548-bdb2-4be5-ba1c-a233442c2bf9 req-e588b14a-79a2-48a0-82b9-c5132deddd65 service nova] Acquiring lock "refresh_cache-6be7d826-e19b-431e-b714-25e646e6b7ca" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.295817] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.408031] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.594877] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Successfully created port: bf459636-3719-4f72-9f4b-e2e9b27811dd {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.630516] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.913368] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Releasing lock "refresh_cache-6be7d826-e19b-431e-b714-25e646e6b7ca" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.913756] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 774.913953] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 774.914284] env[62521]: DEBUG oslo_concurrency.lockutils [req-63c6f548-bdb2-4be5-ba1c-a233442c2bf9 req-e588b14a-79a2-48a0-82b9-c5132deddd65 service nova] Acquired lock "refresh_cache-6be7d826-e19b-431e-b714-25e646e6b7ca" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.914454] env[62521]: DEBUG nova.network.neutron [req-63c6f548-bdb2-4be5-ba1c-a233442c2bf9 req-e588b14a-79a2-48a0-82b9-c5132deddd65 service nova] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Refreshing network info cache for port c0bc58a3-690e-4f33-8b5a-2c55766167f5 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 774.915833] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e0cb4fea-3776-4b7f-8b4c-9a2b1ea4316f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.927494] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6072ef1b-7a36-4585-b413-9dff3fce604d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.955173] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6be7d826-e19b-431e-b714-25e646e6b7ca could not be found. [ 774.955173] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.958582] env[62521]: INFO nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Took 0.04 seconds to destroy the instance on the hypervisor. [ 774.958582] env[62521]: DEBUG oslo.service.loopingcall [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.958772] env[62521]: DEBUG nova.compute.manager [-] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.958772] env[62521]: DEBUG nova.network.neutron [-] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.977694] env[62521]: DEBUG nova.network.neutron [-] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.981976] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8accd50a-1d79-46aa-93b1-4ff0ef2bba6b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.988791] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f107aa99-739e-4411-8e52-da609d06e1d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.020280] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c342e5-a509-4518-ac01-978822db64f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.027865] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ad7a7d-d0c6-4b3f-b8d2-ef19ffcc0963 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.041868] env[62521]: DEBUG nova.compute.provider_tree [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.439121] env[62521]: DEBUG nova.network.neutron [req-63c6f548-bdb2-4be5-ba1c-a233442c2bf9 req-e588b14a-79a2-48a0-82b9-c5132deddd65 service nova] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.475726] env[62521]: ERROR nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bf459636-3719-4f72-9f4b-e2e9b27811dd, please check neutron logs for more information. [ 775.475726] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 775.475726] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.475726] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 775.475726] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.475726] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 775.475726] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.475726] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 775.475726] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.475726] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 775.475726] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.475726] env[62521]: ERROR nova.compute.manager raise self.value [ 775.475726] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.475726] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 775.475726] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.475726] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 775.476526] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.476526] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 775.476526] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bf459636-3719-4f72-9f4b-e2e9b27811dd, please check neutron logs for more information. [ 775.476526] env[62521]: ERROR nova.compute.manager [ 775.476526] env[62521]: Traceback (most recent call last): [ 775.476526] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 775.476526] env[62521]: listener.cb(fileno) [ 775.476526] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.476526] env[62521]: result = function(*args, **kwargs) [ 775.476526] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 775.476526] env[62521]: return func(*args, **kwargs) [ 775.476526] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.476526] env[62521]: raise e [ 775.476526] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.476526] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 775.476526] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.476526] env[62521]: created_port_ids = self._update_ports_for_instance( [ 775.476526] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.476526] env[62521]: with excutils.save_and_reraise_exception(): [ 775.476526] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.476526] env[62521]: self.force_reraise() [ 775.476526] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.476526] env[62521]: raise self.value [ 775.476526] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.476526] env[62521]: updated_port = self._update_port( [ 775.476526] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.476526] env[62521]: _ensure_no_port_binding_failure(port) [ 775.476526] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.476526] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 775.477503] env[62521]: nova.exception.PortBindingFailed: Binding failed for port bf459636-3719-4f72-9f4b-e2e9b27811dd, please check neutron logs for more information. [ 775.477503] env[62521]: Removing descriptor: 18 [ 775.484870] env[62521]: DEBUG nova.network.neutron [-] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.516210] env[62521]: DEBUG nova.network.neutron [req-63c6f548-bdb2-4be5-ba1c-a233442c2bf9 req-e588b14a-79a2-48a0-82b9-c5132deddd65 service nova] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.545181] env[62521]: DEBUG nova.scheduler.client.report [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.645726] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.669781] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.670053] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.670218] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.670409] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.670562] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.670885] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.671142] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.671322] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.671498] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.671660] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.671829] env[62521]: DEBUG nova.virt.hardware [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.672683] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db479ce7-93e7-4994-8cec-7a2750b0611a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.680991] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3014b37-db5f-4cad-a15c-c02a291749a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.697565] env[62521]: ERROR nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bf459636-3719-4f72-9f4b-e2e9b27811dd, please check neutron logs for more information. [ 775.697565] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Traceback (most recent call last): [ 775.697565] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 775.697565] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] yield resources [ 775.697565] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.697565] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] self.driver.spawn(context, instance, image_meta, [ 775.697565] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 775.697565] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.697565] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.697565] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] vm_ref = self.build_virtual_machine(instance, [ 775.697565] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] for vif in network_info: [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] return self._sync_wrapper(fn, *args, **kwargs) [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] self.wait() [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] self[:] = self._gt.wait() [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] return self._exit_event.wait() [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 775.698183] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] current.throw(*self._exc) [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] result = function(*args, **kwargs) [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] return func(*args, **kwargs) [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] raise e [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] nwinfo = self.network_api.allocate_for_instance( [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] created_port_ids = self._update_ports_for_instance( [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] with excutils.save_and_reraise_exception(): [ 775.699728] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] self.force_reraise() [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] raise self.value [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] updated_port = self._update_port( [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] _ensure_no_port_binding_failure(port) [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] raise exception.PortBindingFailed(port_id=port['id']) [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] nova.exception.PortBindingFailed: Binding failed for port bf459636-3719-4f72-9f4b-e2e9b27811dd, please check neutron logs for more information. [ 775.701273] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] [ 775.701273] env[62521]: INFO nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Terminating instance [ 775.703397] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "refresh_cache-590b2441-e627-463b-9722-d0fcb8fcd83a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.703397] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquired lock "refresh_cache-590b2441-e627-463b-9722-d0fcb8fcd83a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.703397] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.986919] env[62521]: INFO nova.compute.manager [-] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Took 1.03 seconds to deallocate network for instance. [ 775.989181] env[62521]: DEBUG nova.compute.claims [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 775.989477] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.018155] env[62521]: DEBUG oslo_concurrency.lockutils [req-63c6f548-bdb2-4be5-ba1c-a233442c2bf9 req-e588b14a-79a2-48a0-82b9-c5132deddd65 service nova] Releasing lock "refresh_cache-6be7d826-e19b-431e-b714-25e646e6b7ca" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.018388] env[62521]: DEBUG nova.compute.manager [req-63c6f548-bdb2-4be5-ba1c-a233442c2bf9 req-e588b14a-79a2-48a0-82b9-c5132deddd65 service nova] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Received event network-vif-deleted-c0bc58a3-690e-4f33-8b5a-2c55766167f5 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.049726] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.050266] env[62521]: DEBUG nova.compute.manager [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 776.053605] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.954s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.055039] env[62521]: INFO nova.compute.claims [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.226406] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.265743] env[62521]: DEBUG nova.compute.manager [req-1d7b74af-c40e-4ee7-b1dd-acf2e58357c9 req-fb2184c5-441b-4a0c-8c8d-605f4209d2b5 service nova] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Received event network-changed-bf459636-3719-4f72-9f4b-e2e9b27811dd {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.265743] env[62521]: DEBUG nova.compute.manager [req-1d7b74af-c40e-4ee7-b1dd-acf2e58357c9 req-fb2184c5-441b-4a0c-8c8d-605f4209d2b5 service nova] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Refreshing instance network info cache due to event network-changed-bf459636-3719-4f72-9f4b-e2e9b27811dd. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.265743] env[62521]: DEBUG oslo_concurrency.lockutils [req-1d7b74af-c40e-4ee7-b1dd-acf2e58357c9 req-fb2184c5-441b-4a0c-8c8d-605f4209d2b5 service nova] Acquiring lock "refresh_cache-590b2441-e627-463b-9722-d0fcb8fcd83a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.337707] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.561050] env[62521]: DEBUG nova.compute.utils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.564477] env[62521]: DEBUG nova.compute.manager [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Not allocating networking since 'none' was specified. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 776.843178] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Releasing lock "refresh_cache-590b2441-e627-463b-9722-d0fcb8fcd83a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.843631] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 776.843827] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 776.844161] env[62521]: DEBUG oslo_concurrency.lockutils [req-1d7b74af-c40e-4ee7-b1dd-acf2e58357c9 req-fb2184c5-441b-4a0c-8c8d-605f4209d2b5 service nova] Acquired lock "refresh_cache-590b2441-e627-463b-9722-d0fcb8fcd83a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.844335] env[62521]: DEBUG nova.network.neutron [req-1d7b74af-c40e-4ee7-b1dd-acf2e58357c9 req-fb2184c5-441b-4a0c-8c8d-605f4209d2b5 service nova] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Refreshing network info cache for port bf459636-3719-4f72-9f4b-e2e9b27811dd {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 776.845455] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-28d8bdda-dcf2-47f9-8e78-2c0cf17851f1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.855054] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-445fa854-dbba-473d-b270-85d5c0b16d67 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.878286] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 590b2441-e627-463b-9722-d0fcb8fcd83a could not be found. [ 776.878521] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 776.878704] env[62521]: INFO nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 776.878943] env[62521]: DEBUG oslo.service.loopingcall [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.879172] env[62521]: DEBUG nova.compute.manager [-] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.879265] env[62521]: DEBUG nova.network.neutron [-] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 776.895130] env[62521]: DEBUG nova.network.neutron [-] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.065255] env[62521]: DEBUG nova.compute.manager [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 777.303894] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8c2764-98ed-4b8f-8fd1-c4b0c29f07b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.311576] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44874ccf-b75c-49c3-9130-c1dad40c8c44 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.341496] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c724445-33ee-46ce-b5ea-0f10bfda53e6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.348535] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9e6312-5d67-4068-9e88-5f0321294a1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.364310] env[62521]: DEBUG nova.compute.provider_tree [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.370830] env[62521]: DEBUG nova.network.neutron [req-1d7b74af-c40e-4ee7-b1dd-acf2e58357c9 req-fb2184c5-441b-4a0c-8c8d-605f4209d2b5 service nova] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.397407] env[62521]: DEBUG nova.network.neutron [-] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.484313] env[62521]: DEBUG nova.network.neutron [req-1d7b74af-c40e-4ee7-b1dd-acf2e58357c9 req-fb2184c5-441b-4a0c-8c8d-605f4209d2b5 service nova] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.869352] env[62521]: DEBUG nova.scheduler.client.report [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.899816] env[62521]: INFO nova.compute.manager [-] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Took 1.02 seconds to deallocate network for instance. [ 777.902142] env[62521]: DEBUG nova.compute.claims [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 777.902778] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.987312] env[62521]: DEBUG oslo_concurrency.lockutils [req-1d7b74af-c40e-4ee7-b1dd-acf2e58357c9 req-fb2184c5-441b-4a0c-8c8d-605f4209d2b5 service nova] Releasing lock "refresh_cache-590b2441-e627-463b-9722-d0fcb8fcd83a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.987661] env[62521]: DEBUG nova.compute.manager [req-1d7b74af-c40e-4ee7-b1dd-acf2e58357c9 req-fb2184c5-441b-4a0c-8c8d-605f4209d2b5 service nova] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Received event network-vif-deleted-bf459636-3719-4f72-9f4b-e2e9b27811dd {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.075840] env[62521]: DEBUG nova.compute.manager [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 778.100900] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.101154] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.101313] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.101493] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.101635] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.101778] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.101979] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.102151] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.102316] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.102477] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.102644] env[62521]: DEBUG nova.virt.hardware [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.103493] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a157bc6d-9bda-4f18-bb6e-38542b593567 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.111128] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f275565-f7b5-48ea-9984-fea36657729f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.124087] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.129496] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Creating folder: Project (435a6e243bc842ae8b9817ee6cf140bc). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.129735] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4fb9572d-0fad-4ffd-b44e-a301a0a47f5b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.140321] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Created folder: Project (435a6e243bc842ae8b9817ee6cf140bc) in parent group-v282025. [ 778.140493] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Creating folder: Instances. Parent ref: group-v282050. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.140686] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3b7a82d9-2d41-40fe-a1a0-d6c351e0edac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.148742] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Created folder: Instances in parent group-v282050. [ 778.148951] env[62521]: DEBUG oslo.service.loopingcall [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.149134] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 778.149310] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7743bf2-b2a1-4351-89a2-273f366b93b1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.164374] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.164374] env[62521]: value = "task-1318429" [ 778.164374] env[62521]: _type = "Task" [ 778.164374] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.171147] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318429, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.374967] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.375638] env[62521]: DEBUG nova.compute.manager [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 778.378135] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.664s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.378318] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.378474] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 778.378754] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.404s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.380528] env[62521]: INFO nova.compute.claims [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.383713] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0619b538-16a8-4234-a370-355ca3779d2d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.391760] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4902d7ef-795b-4c2f-94b3-e8b933178692 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.407609] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cedc47-dc28-448c-b706-e81baf1fe533 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.414566] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa6b85f-7036-4e6d-8b6c-e0cdddc5bc62 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.446086] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181443MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 778.446201] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.674126] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318429, 'name': CreateVM_Task, 'duration_secs': 0.454497} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.674254] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 778.674662] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.674821] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.675173] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 778.675451] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7b37702-0c31-42c2-bb2e-f7059122763b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.679532] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 778.679532] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52292c08-72e0-039c-ccee-3619d53f9433" [ 778.679532] env[62521]: _type = "Task" [ 778.679532] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.686480] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52292c08-72e0-039c-ccee-3619d53f9433, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.884659] env[62521]: DEBUG nova.compute.utils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 778.886131] env[62521]: DEBUG nova.compute.manager [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Not allocating networking since 'none' was specified. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 779.189510] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52292c08-72e0-039c-ccee-3619d53f9433, 'name': SearchDatastore_Task, 'duration_secs': 0.012501} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.189644] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.189832] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.190085] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.190236] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.190417] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.190670] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-11ca43d6-5c5a-4193-b02b-2d239bbc34f7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.198507] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.198633] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 779.199316] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5dd403c9-249d-41bc-a7ea-c904454f5df8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.204009] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 779.204009] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52574000-b1b8-dc29-ec00-9f80db4ab0f4" [ 779.204009] env[62521]: _type = "Task" [ 779.204009] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.211507] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52574000-b1b8-dc29-ec00-9f80db4ab0f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.387468] env[62521]: DEBUG nova.compute.manager [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 779.615384] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e92e567-2a8f-4aaa-880d-ea9803fd527f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.622636] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53fe17b7-cefa-440c-8f1b-2aa692b40035 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.652482] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362bfdef-aba8-446d-84f2-778c0a88480c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.659306] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080a429c-0761-4105-b216-f145807c143e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.671824] env[62521]: DEBUG nova.compute.provider_tree [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.713321] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52574000-b1b8-dc29-ec00-9f80db4ab0f4, 'name': SearchDatastore_Task, 'duration_secs': 0.007964} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.713930] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0b738fd-9889-41d8-9725-ed56dab54917 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.718958] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 779.718958] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b79377-fff0-79db-7dab-8e9ac837f9b4" [ 779.718958] env[62521]: _type = "Task" [ 779.718958] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.726792] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b79377-fff0-79db-7dab-8e9ac837f9b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.174837] env[62521]: DEBUG nova.scheduler.client.report [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.229798] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b79377-fff0-79db-7dab-8e9ac837f9b4, 'name': SearchDatastore_Task, 'duration_secs': 0.009636} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.230066] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.230315] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 062f311c-a783-41cf-8d83-b3a8d4df14f4/062f311c-a783-41cf-8d83-b3a8d4df14f4.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 780.230562] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-321df71e-1894-4a0a-80e5-28862477ca5f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.236827] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 780.236827] env[62521]: value = "task-1318430" [ 780.236827] env[62521]: _type = "Task" [ 780.236827] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.243946] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318430, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.399071] env[62521]: DEBUG nova.compute.manager [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 780.420347] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 780.420633] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 780.420796] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.420978] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 780.421141] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.421290] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 780.421518] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 780.421679] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 780.421843] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 780.422020] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 780.422213] env[62521]: DEBUG nova.virt.hardware [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.423091] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbdebb59-6daa-4c92-8a87-51d10fa4abea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.433178] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1e199d-c991-473a-930e-c31f809fa6dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.447173] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 780.453669] env[62521]: DEBUG oslo.service.loopingcall [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.454447] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 780.454728] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37f07c34-4304-41c8-9654-7fbb3ae44984 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.473695] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.473695] env[62521]: value = "task-1318431" [ 780.473695] env[62521]: _type = "Task" [ 780.473695] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.482292] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318431, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.680054] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.301s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.680545] env[62521]: DEBUG nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 780.683568] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.058s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.746112] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318430, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485644} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.746400] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 062f311c-a783-41cf-8d83-b3a8d4df14f4/062f311c-a783-41cf-8d83-b3a8d4df14f4.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 780.746622] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.746866] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a57f99a2-a1d7-43de-aa25-70734816f4b8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.753572] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 780.753572] env[62521]: value = "task-1318432" [ 780.753572] env[62521]: _type = "Task" [ 780.753572] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.761129] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318432, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.983548] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318431, 'name': CreateVM_Task, 'duration_secs': 0.369154} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.983715] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 780.984139] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.984301] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.984624] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 780.984865] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2512466f-8e71-4a05-9d41-10997c788550 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.988957] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 780.988957] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527faa6a-0fd9-380e-e4b0-59ab91d9c101" [ 780.988957] env[62521]: _type = "Task" [ 780.988957] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.996197] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527faa6a-0fd9-380e-e4b0-59ab91d9c101, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.188910] env[62521]: DEBUG nova.compute.utils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 781.193377] env[62521]: DEBUG nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 781.193660] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 781.245309] env[62521]: DEBUG nova.policy [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb922c81aff243848d13512dd121671c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f79a1bc2197494e83fedcddaa26d3a2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 781.262868] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318432, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070771} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.263250] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.264051] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8545742b-f880-46ee-a716-478a56582b72 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.285555] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 062f311c-a783-41cf-8d83-b3a8d4df14f4/062f311c-a783-41cf-8d83-b3a8d4df14f4.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.287405] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ada95ce7-721b-4065-b8b1-eb3ab5958d2e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.308205] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 781.308205] env[62521]: value = "task-1318433" [ 781.308205] env[62521]: _type = "Task" [ 781.308205] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.318314] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318433, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.458900] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a98265-d8d0-4d24-a799-73f496b61063 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.468399] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94f461a-75e0-48a9-bd41-ecbc240724ec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.500831] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2f5a86-5079-4a8b-bae6-e7c1f014d0a4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.509437] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527faa6a-0fd9-380e-e4b0-59ab91d9c101, 'name': SearchDatastore_Task, 'duration_secs': 0.00924} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.511622] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.511876] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.512119] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.512269] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.512464] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.512750] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-68f51f21-72e2-45c4-a940-fd9843d0da8f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.515466] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be73293-9297-4459-81cb-11fbadb10cee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.528909] env[62521]: DEBUG nova.compute.provider_tree [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.531102] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.531294] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 781.532218] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c547bec8-f3ff-4846-9e01-c1fb1b0f3b71 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.537132] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 781.537132] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525fdc61-8f13-37db-edb6-4b52d0b96e3f" [ 781.537132] env[62521]: _type = "Task" [ 781.537132] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.544830] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525fdc61-8f13-37db-edb6-4b52d0b96e3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.586396] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Successfully created port: 5e83db33-8f13-4a83-b35d-e9811b71dc4a {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.694383] env[62521]: DEBUG nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.818017] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318433, 'name': ReconfigVM_Task, 'duration_secs': 0.325012} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.818302] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 062f311c-a783-41cf-8d83-b3a8d4df14f4/062f311c-a783-41cf-8d83-b3a8d4df14f4.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.818972] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-421c2557-be01-4440-b7f8-7ca5252714c0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.825082] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 781.825082] env[62521]: value = "task-1318434" [ 781.825082] env[62521]: _type = "Task" [ 781.825082] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.837051] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318434, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.033863] env[62521]: DEBUG nova.scheduler.client.report [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.046965] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525fdc61-8f13-37db-edb6-4b52d0b96e3f, 'name': SearchDatastore_Task, 'duration_secs': 0.009168} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.048116] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-666d0e2e-a12a-4e42-95e6-12d276d73d3b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.053978] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 782.053978] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524938fe-a4f5-f114-7f0e-5ce037baf26d" [ 782.053978] env[62521]: _type = "Task" [ 782.053978] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.062221] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524938fe-a4f5-f114-7f0e-5ce037baf26d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.334671] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318434, 'name': Rename_Task, 'duration_secs': 0.137404} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.334986] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.335378] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6340ff48-9d5d-4168-9629-08d4f0034d94 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.341756] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 782.341756] env[62521]: value = "task-1318435" [ 782.341756] env[62521]: _type = "Task" [ 782.341756] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.354182] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318435, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.538968] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.855s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.539639] env[62521]: ERROR nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 16a811d0-87b7-45a0-b083-cb9a515738c4, please check neutron logs for more information. [ 782.539639] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Traceback (most recent call last): [ 782.539639] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 782.539639] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] self.driver.spawn(context, instance, image_meta, [ 782.539639] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 782.539639] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 782.539639] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 782.539639] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] vm_ref = self.build_virtual_machine(instance, [ 782.539639] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 782.539639] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] vif_infos = vmwarevif.get_vif_info(self._session, [ 782.539639] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] for vif in network_info: [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] return self._sync_wrapper(fn, *args, **kwargs) [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] self.wait() [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] self[:] = self._gt.wait() [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] return self._exit_event.wait() [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] result = hub.switch() [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 782.540045] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] return self.greenlet.switch() [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] result = function(*args, **kwargs) [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] return func(*args, **kwargs) [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] raise e [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] nwinfo = self.network_api.allocate_for_instance( [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] created_port_ids = self._update_ports_for_instance( [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] with excutils.save_and_reraise_exception(): [ 782.540474] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] self.force_reraise() [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] raise self.value [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] updated_port = self._update_port( [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] _ensure_no_port_binding_failure(port) [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] raise exception.PortBindingFailed(port_id=port['id']) [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] nova.exception.PortBindingFailed: Binding failed for port 16a811d0-87b7-45a0-b083-cb9a515738c4, please check neutron logs for more information. [ 782.540901] env[62521]: ERROR nova.compute.manager [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] [ 782.541298] env[62521]: DEBUG nova.compute.utils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Binding failed for port 16a811d0-87b7-45a0-b083-cb9a515738c4, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 782.541606] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.705s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.544465] env[62521]: DEBUG nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Build of instance 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a was re-scheduled: Binding failed for port 16a811d0-87b7-45a0-b083-cb9a515738c4, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 782.544912] env[62521]: DEBUG nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 782.545155] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Acquiring lock "refresh_cache-0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.545303] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Acquired lock "refresh_cache-0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.545518] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.564314] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524938fe-a4f5-f114-7f0e-5ce037baf26d, 'name': SearchDatastore_Task, 'duration_secs': 0.009137} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.564558] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.564804] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 868c7798-8d0f-4c31-88dc-5007454796f1/868c7798-8d0f-4c31-88dc-5007454796f1.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 782.565065] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13129a32-2532-4ba6-ae03-47ca71a46e9a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.571536] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 782.571536] env[62521]: value = "task-1318436" [ 782.571536] env[62521]: _type = "Task" [ 782.571536] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.580152] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318436, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.627145] env[62521]: DEBUG nova.compute.manager [req-9c70da0f-d641-45fc-b896-31928ba946f5 req-97ed3e9c-b9f6-438c-a3f4-798c01489e18 service nova] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Received event network-changed-5e83db33-8f13-4a83-b35d-e9811b71dc4a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.627384] env[62521]: DEBUG nova.compute.manager [req-9c70da0f-d641-45fc-b896-31928ba946f5 req-97ed3e9c-b9f6-438c-a3f4-798c01489e18 service nova] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Refreshing instance network info cache due to event network-changed-5e83db33-8f13-4a83-b35d-e9811b71dc4a. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.627613] env[62521]: DEBUG oslo_concurrency.lockutils [req-9c70da0f-d641-45fc-b896-31928ba946f5 req-97ed3e9c-b9f6-438c-a3f4-798c01489e18 service nova] Acquiring lock "refresh_cache-de583c0b-d147-4f54-bc90-49bdd71c56aa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.627786] env[62521]: DEBUG oslo_concurrency.lockutils [req-9c70da0f-d641-45fc-b896-31928ba946f5 req-97ed3e9c-b9f6-438c-a3f4-798c01489e18 service nova] Acquired lock "refresh_cache-de583c0b-d147-4f54-bc90-49bdd71c56aa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.627948] env[62521]: DEBUG nova.network.neutron [req-9c70da0f-d641-45fc-b896-31928ba946f5 req-97ed3e9c-b9f6-438c-a3f4-798c01489e18 service nova] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Refreshing network info cache for port 5e83db33-8f13-4a83-b35d-e9811b71dc4a {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 782.704096] env[62521]: DEBUG nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.733199] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.733731] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.733731] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.733964] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.734067] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.734208] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.734451] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.734653] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.734831] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.734996] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.735186] env[62521]: DEBUG nova.virt.hardware [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.736123] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff6cdf4-a62f-4ec4-b0f1-beefd1c059f1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.744676] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162b11ea-f534-44cb-8f19-ba5692509d8f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.790969] env[62521]: ERROR nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5e83db33-8f13-4a83-b35d-e9811b71dc4a, please check neutron logs for more information. [ 782.790969] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 782.790969] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.790969] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 782.790969] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.790969] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 782.790969] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.790969] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 782.790969] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.790969] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 782.790969] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.790969] env[62521]: ERROR nova.compute.manager raise self.value [ 782.790969] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.790969] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 782.790969] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.790969] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 782.791558] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.791558] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 782.791558] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5e83db33-8f13-4a83-b35d-e9811b71dc4a, please check neutron logs for more information. [ 782.791558] env[62521]: ERROR nova.compute.manager [ 782.791558] env[62521]: Traceback (most recent call last): [ 782.791558] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 782.791558] env[62521]: listener.cb(fileno) [ 782.791558] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.791558] env[62521]: result = function(*args, **kwargs) [ 782.791558] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 782.791558] env[62521]: return func(*args, **kwargs) [ 782.791558] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.791558] env[62521]: raise e [ 782.791558] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.791558] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 782.791558] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.791558] env[62521]: created_port_ids = self._update_ports_for_instance( [ 782.791558] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.791558] env[62521]: with excutils.save_and_reraise_exception(): [ 782.791558] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.791558] env[62521]: self.force_reraise() [ 782.791558] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.791558] env[62521]: raise self.value [ 782.791558] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.791558] env[62521]: updated_port = self._update_port( [ 782.791558] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.791558] env[62521]: _ensure_no_port_binding_failure(port) [ 782.791558] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.791558] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 782.792590] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 5e83db33-8f13-4a83-b35d-e9811b71dc4a, please check neutron logs for more information. [ 782.792590] env[62521]: Removing descriptor: 15 [ 782.792590] env[62521]: ERROR nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5e83db33-8f13-4a83-b35d-e9811b71dc4a, please check neutron logs for more information. [ 782.792590] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Traceback (most recent call last): [ 782.792590] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 782.792590] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] yield resources [ 782.792590] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 782.792590] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] self.driver.spawn(context, instance, image_meta, [ 782.792590] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 782.792590] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 782.792590] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 782.792590] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] vm_ref = self.build_virtual_machine(instance, [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] vif_infos = vmwarevif.get_vif_info(self._session, [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] for vif in network_info: [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] return self._sync_wrapper(fn, *args, **kwargs) [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] self.wait() [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] self[:] = self._gt.wait() [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] return self._exit_event.wait() [ 782.793195] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] result = hub.switch() [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] return self.greenlet.switch() [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] result = function(*args, **kwargs) [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] return func(*args, **kwargs) [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] raise e [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] nwinfo = self.network_api.allocate_for_instance( [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.793691] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] created_port_ids = self._update_ports_for_instance( [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] with excutils.save_and_reraise_exception(): [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] self.force_reraise() [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] raise self.value [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] updated_port = self._update_port( [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] _ensure_no_port_binding_failure(port) [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.794169] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] raise exception.PortBindingFailed(port_id=port['id']) [ 782.794750] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] nova.exception.PortBindingFailed: Binding failed for port 5e83db33-8f13-4a83-b35d-e9811b71dc4a, please check neutron logs for more information. [ 782.794750] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] [ 782.794750] env[62521]: INFO nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Terminating instance [ 782.795679] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Acquiring lock "refresh_cache-de583c0b-d147-4f54-bc90-49bdd71c56aa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.852710] env[62521]: DEBUG oslo_vmware.api [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318435, 'name': PowerOnVM_Task, 'duration_secs': 0.434265} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.853017] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 782.853237] env[62521]: INFO nova.compute.manager [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Took 4.78 seconds to spawn the instance on the hypervisor. [ 782.853456] env[62521]: DEBUG nova.compute.manager [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.854297] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7afedf4-c83a-463f-8dbb-1cb540048d06 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.087602] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.089353] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318436, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509526} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.089791] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 868c7798-8d0f-4c31-88dc-5007454796f1/868c7798-8d0f-4c31-88dc-5007454796f1.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 783.090017] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.091457] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c401a7c-b181-45c0-8903-c84cad8c685e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.102015] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 783.102015] env[62521]: value = "task-1318437" [ 783.102015] env[62521]: _type = "Task" [ 783.102015] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.114926] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318437, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.152311] env[62521]: DEBUG nova.network.neutron [req-9c70da0f-d641-45fc-b896-31928ba946f5 req-97ed3e9c-b9f6-438c-a3f4-798c01489e18 service nova] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.245991] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.262347] env[62521]: DEBUG nova.network.neutron [req-9c70da0f-d641-45fc-b896-31928ba946f5 req-97ed3e9c-b9f6-438c-a3f4-798c01489e18 service nova] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.372767] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343e4ad6-489c-4a88-8de5-49bff23e5424 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.377619] env[62521]: INFO nova.compute.manager [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Took 25.21 seconds to build instance. [ 783.384017] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d46f0c-5fff-4224-87ce-e0af515b2fbf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.412542] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ed4f4d-fde1-4521-a622-1df00d1efc6b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.420014] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5767dcf-e64d-44eb-97a6-0e66666946c8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.434493] env[62521]: DEBUG nova.compute.provider_tree [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.611916] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318437, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06303} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.613022] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.613251] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b5e8d8-8f49-434d-9046-c985261bf20b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.632236] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 868c7798-8d0f-4c31-88dc-5007454796f1/868c7798-8d0f-4c31-88dc-5007454796f1.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.632479] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e73ded7-1340-4c1b-b409-c5d247fde233 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.651046] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 783.651046] env[62521]: value = "task-1318438" [ 783.651046] env[62521]: _type = "Task" [ 783.651046] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.662674] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318438, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.748957] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Releasing lock "refresh_cache-0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.749253] env[62521]: DEBUG nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 783.749441] env[62521]: DEBUG nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.749613] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 783.764323] env[62521]: DEBUG oslo_concurrency.lockutils [req-9c70da0f-d641-45fc-b896-31928ba946f5 req-97ed3e9c-b9f6-438c-a3f4-798c01489e18 service nova] Releasing lock "refresh_cache-de583c0b-d147-4f54-bc90-49bdd71c56aa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.764706] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Acquired lock "refresh_cache-de583c0b-d147-4f54-bc90-49bdd71c56aa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.764887] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.767674] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.880363] env[62521]: DEBUG oslo_concurrency.lockutils [None req-897009c7-1b70-4cc2-b702-a092afb0cae2 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "062f311c-a783-41cf-8d83-b3a8d4df14f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.270s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.937948] env[62521]: DEBUG nova.scheduler.client.report [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.161070] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.271871] env[62521]: DEBUG nova.network.neutron [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.282655] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.352587] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.383541] env[62521]: DEBUG nova.compute.manager [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.442447] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.901s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.444029] env[62521]: ERROR nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4365d026-b662-4559-b65f-4d64ad164f6e, please check neutron logs for more information. [ 784.444029] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Traceback (most recent call last): [ 784.444029] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 784.444029] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] self.driver.spawn(context, instance, image_meta, [ 784.444029] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 784.444029] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] self._vmops.spawn(context, instance, image_meta, injected_files, [ 784.444029] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 784.444029] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] vm_ref = self.build_virtual_machine(instance, [ 784.444029] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 784.444029] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] vif_infos = vmwarevif.get_vif_info(self._session, [ 784.444029] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] for vif in network_info: [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] return self._sync_wrapper(fn, *args, **kwargs) [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] self.wait() [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] self[:] = self._gt.wait() [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] return self._exit_event.wait() [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] result = hub.switch() [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 784.444462] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] return self.greenlet.switch() [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] result = function(*args, **kwargs) [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] return func(*args, **kwargs) [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] raise e [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] nwinfo = self.network_api.allocate_for_instance( [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] created_port_ids = self._update_ports_for_instance( [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] with excutils.save_and_reraise_exception(): [ 784.444862] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] self.force_reraise() [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] raise self.value [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] updated_port = self._update_port( [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] _ensure_no_port_binding_failure(port) [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] raise exception.PortBindingFailed(port_id=port['id']) [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] nova.exception.PortBindingFailed: Binding failed for port 4365d026-b662-4559-b65f-4d64ad164f6e, please check neutron logs for more information. [ 784.445294] env[62521]: ERROR nova.compute.manager [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] [ 784.445688] env[62521]: DEBUG nova.compute.utils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Binding failed for port 4365d026-b662-4559-b65f-4d64ad164f6e, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 784.445688] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.821s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.447447] env[62521]: INFO nova.compute.claims [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 784.450830] env[62521]: DEBUG nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Build of instance 32c0268b-eee6-4e45-b9c3-077ec81bd314 was re-scheduled: Binding failed for port 4365d026-b662-4559-b65f-4d64ad164f6e, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 784.451307] env[62521]: DEBUG nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 784.451542] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Acquiring lock "refresh_cache-32c0268b-eee6-4e45-b9c3-077ec81bd314" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.451690] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Acquired lock "refresh_cache-32c0268b-eee6-4e45-b9c3-077ec81bd314" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.451846] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 784.659438] env[62521]: DEBUG nova.compute.manager [req-78646b95-33f7-4cf8-a28f-2ab6f1664d3e req-a24b8525-568a-427e-82ec-f345e52d52cd service nova] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Received event network-vif-deleted-5e83db33-8f13-4a83-b35d-e9811b71dc4a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.665387] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318438, 'name': ReconfigVM_Task, 'duration_secs': 0.991806} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.665633] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 868c7798-8d0f-4c31-88dc-5007454796f1/868c7798-8d0f-4c31-88dc-5007454796f1.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.666202] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ecd78c6-1b71-4063-a553-68b3cdd826fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.671903] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 784.671903] env[62521]: value = "task-1318439" [ 784.671903] env[62521]: _type = "Task" [ 784.671903] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.679596] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318439, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.774885] env[62521]: INFO nova.compute.manager [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] [instance: 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a] Took 1.02 seconds to deallocate network for instance. [ 784.855340] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Releasing lock "refresh_cache-de583c0b-d147-4f54-bc90-49bdd71c56aa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.855901] env[62521]: DEBUG nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 784.856212] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 784.856592] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-734993b1-3fb0-44bd-bcd7-531cdc7cc8b1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.868032] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ef1f6c-4cfa-441d-8f8d-1e6b07dd7011 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.900310] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance de583c0b-d147-4f54-bc90-49bdd71c56aa could not be found. [ 784.900617] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.900880] env[62521]: INFO nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Took 0.04 seconds to destroy the instance on the hypervisor. [ 784.901259] env[62521]: DEBUG oslo.service.loopingcall [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.901570] env[62521]: DEBUG nova.compute.manager [-] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.901793] env[62521]: DEBUG nova.network.neutron [-] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.904518] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.917740] env[62521]: DEBUG nova.network.neutron [-] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.967903] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.038643] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.181771] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318439, 'name': Rename_Task, 'duration_secs': 0.131556} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.182069] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 785.182308] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb175800-9c01-410c-b36a-866dca662eaa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.188683] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 785.188683] env[62521]: value = "task-1318440" [ 785.188683] env[62521]: _type = "Task" [ 785.188683] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.195916] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318440, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.420556] env[62521]: DEBUG nova.network.neutron [-] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.542260] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Releasing lock "refresh_cache-32c0268b-eee6-4e45-b9c3-077ec81bd314" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.542534] env[62521]: DEBUG nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 785.542698] env[62521]: DEBUG nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.542978] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 785.561322] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.703138] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318440, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.719020] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc3b6a9-4c28-4790-8f8d-648b57952c8d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.724840] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a91131e-103a-4d33-8ec2-f5a9d1a2bbbf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.756738] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5580263d-72ee-4a63-a657-d3aaa6a163c8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.764739] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ed7106-0ce6-47ba-bdee-89345bb48f18 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.778617] env[62521]: DEBUG nova.compute.provider_tree [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.803471] env[62521]: INFO nova.scheduler.client.report [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Deleted allocations for instance 0b6fb3f8-07ce-458c-9efe-33ea85bbe62a [ 785.923584] env[62521]: INFO nova.compute.manager [-] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Took 1.02 seconds to deallocate network for instance. [ 785.926818] env[62521]: DEBUG nova.compute.claims [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 785.927070] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.064804] env[62521]: DEBUG nova.network.neutron [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.199619] env[62521]: DEBUG oslo_vmware.api [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318440, 'name': PowerOnVM_Task, 'duration_secs': 0.807746} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.199884] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 786.200101] env[62521]: INFO nova.compute.manager [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Took 5.80 seconds to spawn the instance on the hypervisor. [ 786.200280] env[62521]: DEBUG nova.compute.manager [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 786.201032] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63846b58-0d65-4338-b15e-5a567acb01cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.282030] env[62521]: DEBUG nova.scheduler.client.report [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.311319] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02d05c33-7cd3-44dc-8def-2de9004f0802 tempest-AttachInterfacesUnderV243Test-475351532 tempest-AttachInterfacesUnderV243Test-475351532-project-member] Lock "0b6fb3f8-07ce-458c-9efe-33ea85bbe62a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 172.293s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.567412] env[62521]: INFO nova.compute.manager [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] [instance: 32c0268b-eee6-4e45-b9c3-077ec81bd314] Took 1.02 seconds to deallocate network for instance. [ 786.715413] env[62521]: INFO nova.compute.manager [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Took 26.63 seconds to build instance. [ 786.787728] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.788259] env[62521]: DEBUG nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 786.791477] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.245s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.814090] env[62521]: DEBUG nova.compute.manager [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 787.218163] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7440585c-8075-4e9b-9d97-cd5f076c63ff tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "868c7798-8d0f-4c31-88dc-5007454796f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.204s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.299768] env[62521]: DEBUG nova.compute.utils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.303682] env[62521]: DEBUG nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 787.303682] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 787.339337] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.366944] env[62521]: DEBUG nova.policy [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '100eb6333b6849569e15158f56df7830', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e5f737d9342a422cbded80b1d6a8d0b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.608426] env[62521]: INFO nova.scheduler.client.report [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Deleted allocations for instance 32c0268b-eee6-4e45-b9c3-077ec81bd314 [ 787.616800] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4318639-ed81-4ffa-b7d3-6fc52935d870 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.625388] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b83a3f-179e-47a3-a578-c90651b19dad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.662502] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577dca48-c17d-4f1d-a54f-6af475e318fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.670430] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6015a5f1-9805-4865-b557-9cfec18f2844 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.685168] env[62521]: DEBUG nova.compute.provider_tree [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.719444] env[62521]: DEBUG nova.compute.manager [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 787.806867] env[62521]: DEBUG nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.908018] env[62521]: INFO nova.compute.manager [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Rebuilding instance [ 787.952964] env[62521]: DEBUG nova.compute.manager [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 787.953956] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c6a28a-f650-4538-9cfb-16db36c6c1f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.963548] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Successfully created port: 3d591a87-7f16-445c-80f4-c1f04dafd7b6 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.122527] env[62521]: DEBUG oslo_concurrency.lockutils [None req-03ee737f-0d0a-4503-9db9-30cb7e8b95d2 tempest-ServerGroupTestJSON-1192523842 tempest-ServerGroupTestJSON-1192523842-project-member] Lock "32c0268b-eee6-4e45-b9c3-077ec81bd314" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 170.318s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.188554] env[62521]: DEBUG nova.scheduler.client.report [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.239501] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.465239] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 788.465874] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c737be1-1b09-4218-b96f-b7081f2681ab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.477529] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 788.477529] env[62521]: value = "task-1318441" [ 788.477529] env[62521]: _type = "Task" [ 788.477529] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.497967] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318441, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.627266] env[62521]: DEBUG nova.compute.manager [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.699868] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.908s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.700524] env[62521]: ERROR nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4bee5029-6582-4a51-9759-3d91534d13b5, please check neutron logs for more information. [ 788.700524] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Traceback (most recent call last): [ 788.700524] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 788.700524] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] self.driver.spawn(context, instance, image_meta, [ 788.700524] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 788.700524] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 788.700524] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 788.700524] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] vm_ref = self.build_virtual_machine(instance, [ 788.700524] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 788.700524] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 788.700524] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] for vif in network_info: [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] return self._sync_wrapper(fn, *args, **kwargs) [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] self.wait() [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] self[:] = self._gt.wait() [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] return self._exit_event.wait() [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] current.throw(*self._exc) [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 788.700907] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] result = function(*args, **kwargs) [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] return func(*args, **kwargs) [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] raise e [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] nwinfo = self.network_api.allocate_for_instance( [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] created_port_ids = self._update_ports_for_instance( [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] with excutils.save_and_reraise_exception(): [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] self.force_reraise() [ 788.701300] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.701720] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] raise self.value [ 788.701720] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 788.701720] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] updated_port = self._update_port( [ 788.701720] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.701720] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] _ensure_no_port_binding_failure(port) [ 788.701720] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.701720] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] raise exception.PortBindingFailed(port_id=port['id']) [ 788.701720] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] nova.exception.PortBindingFailed: Binding failed for port 4bee5029-6582-4a51-9759-3d91534d13b5, please check neutron logs for more information. [ 788.701720] env[62521]: ERROR nova.compute.manager [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] [ 788.702325] env[62521]: DEBUG nova.compute.utils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Binding failed for port 4bee5029-6582-4a51-9759-3d91534d13b5, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 788.703688] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.698s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.708451] env[62521]: DEBUG nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Build of instance 63abbe0a-af36-4679-8cf1-87f9e9abf7b9 was re-scheduled: Binding failed for port 4bee5029-6582-4a51-9759-3d91534d13b5, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 788.711980] env[62521]: DEBUG nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 788.711980] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Acquiring lock "refresh_cache-63abbe0a-af36-4679-8cf1-87f9e9abf7b9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.711980] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Acquired lock "refresh_cache-63abbe0a-af36-4679-8cf1-87f9e9abf7b9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.711980] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.816054] env[62521]: DEBUG nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.841212] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.841474] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.841636] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.841821] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.841967] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.842471] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.842722] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.842888] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.843068] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.843233] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.843439] env[62521]: DEBUG nova.virt.hardware [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.845415] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd65aae4-93d4-4a92-9c21-c872f68ac4e0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.853412] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0ac658-ab35-4c2d-8395-bb914e849477 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.990369] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318441, 'name': PowerOffVM_Task, 'duration_secs': 0.136249} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.990679] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 788.990908] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 788.991697] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6fd9aa-8726-4cdf-beed-4ad7f1fdcefa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.002666] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 789.002917] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ded487a9-19e6-4c58-a566-8a5e1228e934 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.030197] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 789.030422] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 789.030601] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Deleting the datastore file [datastore1] 868c7798-8d0f-4c31-88dc-5007454796f1 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 789.030860] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3285869f-e203-4dd4-ac9e-8436447833c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.036386] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 789.036386] env[62521]: value = "task-1318443" [ 789.036386] env[62521]: _type = "Task" [ 789.036386] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.045886] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318443, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.112014] env[62521]: DEBUG nova.compute.manager [req-bcda3bce-ac1d-4ae8-8525-a9bbbc95f752 req-608c422c-eb40-4f89-aee2-51052a79cc48 service nova] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Received event network-changed-3d591a87-7f16-445c-80f4-c1f04dafd7b6 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.112014] env[62521]: DEBUG nova.compute.manager [req-bcda3bce-ac1d-4ae8-8525-a9bbbc95f752 req-608c422c-eb40-4f89-aee2-51052a79cc48 service nova] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Refreshing instance network info cache due to event network-changed-3d591a87-7f16-445c-80f4-c1f04dafd7b6. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 789.112014] env[62521]: DEBUG oslo_concurrency.lockutils [req-bcda3bce-ac1d-4ae8-8525-a9bbbc95f752 req-608c422c-eb40-4f89-aee2-51052a79cc48 service nova] Acquiring lock "refresh_cache-3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.112014] env[62521]: DEBUG oslo_concurrency.lockutils [req-bcda3bce-ac1d-4ae8-8525-a9bbbc95f752 req-608c422c-eb40-4f89-aee2-51052a79cc48 service nova] Acquired lock "refresh_cache-3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.112014] env[62521]: DEBUG nova.network.neutron [req-bcda3bce-ac1d-4ae8-8525-a9bbbc95f752 req-608c422c-eb40-4f89-aee2-51052a79cc48 service nova] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Refreshing network info cache for port 3d591a87-7f16-445c-80f4-c1f04dafd7b6 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 789.157875] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.241779] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.412979] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.423775] env[62521]: ERROR nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3d591a87-7f16-445c-80f4-c1f04dafd7b6, please check neutron logs for more information. [ 789.423775] env[62521]: ERROR nova.compute.manager Traceback (most recent call last): [ 789.423775] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.423775] env[62521]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 789.423775] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.423775] env[62521]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 789.423775] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.423775] env[62521]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 789.423775] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.423775] env[62521]: ERROR nova.compute.manager self.force_reraise() [ 789.423775] env[62521]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.423775] env[62521]: ERROR nova.compute.manager raise self.value [ 789.423775] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.423775] env[62521]: ERROR nova.compute.manager updated_port = self._update_port( [ 789.423775] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.423775] env[62521]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 789.424363] env[62521]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.424363] env[62521]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 789.424363] env[62521]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3d591a87-7f16-445c-80f4-c1f04dafd7b6, please check neutron logs for more information. [ 789.424363] env[62521]: ERROR nova.compute.manager [ 789.424363] env[62521]: Traceback (most recent call last): [ 789.424363] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 789.424363] env[62521]: listener.cb(fileno) [ 789.424363] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.424363] env[62521]: result = function(*args, **kwargs) [ 789.424363] env[62521]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.424363] env[62521]: return func(*args, **kwargs) [ 789.424363] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.424363] env[62521]: raise e [ 789.424363] env[62521]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.424363] env[62521]: nwinfo = self.network_api.allocate_for_instance( [ 789.424363] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.424363] env[62521]: created_port_ids = self._update_ports_for_instance( [ 789.424363] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.424363] env[62521]: with excutils.save_and_reraise_exception(): [ 789.424363] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.424363] env[62521]: self.force_reraise() [ 789.424363] env[62521]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.424363] env[62521]: raise self.value [ 789.424363] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.424363] env[62521]: updated_port = self._update_port( [ 789.424363] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.424363] env[62521]: _ensure_no_port_binding_failure(port) [ 789.424363] env[62521]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.424363] env[62521]: raise exception.PortBindingFailed(port_id=port['id']) [ 789.425321] env[62521]: nova.exception.PortBindingFailed: Binding failed for port 3d591a87-7f16-445c-80f4-c1f04dafd7b6, please check neutron logs for more information. [ 789.425321] env[62521]: Removing descriptor: 18 [ 789.428115] env[62521]: ERROR nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3d591a87-7f16-445c-80f4-c1f04dafd7b6, please check neutron logs for more information. [ 789.428115] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Traceback (most recent call last): [ 789.428115] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 789.428115] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] yield resources [ 789.428115] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 789.428115] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] self.driver.spawn(context, instance, image_meta, [ 789.428115] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 789.428115] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 789.428115] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 789.428115] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] vm_ref = self.build_virtual_machine(instance, [ 789.428115] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] vif_infos = vmwarevif.get_vif_info(self._session, [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] for vif in network_info: [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] return self._sync_wrapper(fn, *args, **kwargs) [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] self.wait() [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] self[:] = self._gt.wait() [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] return self._exit_event.wait() [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 789.429931] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] result = hub.switch() [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] return self.greenlet.switch() [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] result = function(*args, **kwargs) [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] return func(*args, **kwargs) [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] raise e [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] nwinfo = self.network_api.allocate_for_instance( [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] created_port_ids = self._update_ports_for_instance( [ 789.430415] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] with excutils.save_and_reraise_exception(): [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] self.force_reraise() [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] raise self.value [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] updated_port = self._update_port( [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] _ensure_no_port_binding_failure(port) [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] raise exception.PortBindingFailed(port_id=port['id']) [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] nova.exception.PortBindingFailed: Binding failed for port 3d591a87-7f16-445c-80f4-c1f04dafd7b6, please check neutron logs for more information. [ 789.430866] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] [ 789.431367] env[62521]: INFO nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Terminating instance [ 789.431367] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Acquiring lock "refresh_cache-3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.501230] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2ad784-5fcb-4256-9bc7-451ca69b5850 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.509258] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f467af9f-5ad1-4bdb-ae1e-e3bdaaf12dcc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.547038] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cacded6f-8084-4daa-bdbc-d6fc5e35833f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.555324] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318443, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108189} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.559964] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.560193] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 789.560376] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.567708] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e9bc3b-123c-4edb-82c9-5a3d30a18f35 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.584302] env[62521]: DEBUG nova.compute.provider_tree [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.634851] env[62521]: DEBUG nova.network.neutron [req-bcda3bce-ac1d-4ae8-8525-a9bbbc95f752 req-608c422c-eb40-4f89-aee2-51052a79cc48 service nova] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.754467] env[62521]: DEBUG nova.network.neutron [req-bcda3bce-ac1d-4ae8-8525-a9bbbc95f752 req-608c422c-eb40-4f89-aee2-51052a79cc48 service nova] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.917190] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Releasing lock "refresh_cache-63abbe0a-af36-4679-8cf1-87f9e9abf7b9" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.917449] env[62521]: DEBUG nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 789.917642] env[62521]: DEBUG nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.917807] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 789.941886] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.087600] env[62521]: DEBUG nova.scheduler.client.report [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.256990] env[62521]: DEBUG oslo_concurrency.lockutils [req-bcda3bce-ac1d-4ae8-8525-a9bbbc95f752 req-608c422c-eb40-4f89-aee2-51052a79cc48 service nova] Releasing lock "refresh_cache-3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.257580] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Acquired lock "refresh_cache-3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.258577] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.446373] env[62521]: DEBUG nova.network.neutron [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.593709] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.890s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.594404] env[62521]: ERROR nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eaa99398-192b-4f83-a3fb-a0ac0164fa23, please check neutron logs for more information. [ 790.594404] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Traceback (most recent call last): [ 790.594404] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 790.594404] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] self.driver.spawn(context, instance, image_meta, [ 790.594404] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 790.594404] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.594404] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.594404] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] vm_ref = self.build_virtual_machine(instance, [ 790.594404] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 790.594404] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] vif_infos = vmwarevif.get_vif_info(self._session, [ 790.594404] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] for vif in network_info: [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] return self._sync_wrapper(fn, *args, **kwargs) [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] self.wait() [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] self[:] = self._gt.wait() [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] return self._exit_event.wait() [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] current.throw(*self._exc) [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.594765] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] result = function(*args, **kwargs) [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] return func(*args, **kwargs) [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] raise e [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] nwinfo = self.network_api.allocate_for_instance( [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] created_port_ids = self._update_ports_for_instance( [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] with excutils.save_and_reraise_exception(): [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] self.force_reraise() [ 790.595153] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.595595] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] raise self.value [ 790.595595] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.595595] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] updated_port = self._update_port( [ 790.595595] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.595595] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] _ensure_no_port_binding_failure(port) [ 790.595595] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.595595] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] raise exception.PortBindingFailed(port_id=port['id']) [ 790.595595] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] nova.exception.PortBindingFailed: Binding failed for port eaa99398-192b-4f83-a3fb-a0ac0164fa23, please check neutron logs for more information. [ 790.595595] env[62521]: ERROR nova.compute.manager [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] [ 790.595595] env[62521]: DEBUG nova.compute.utils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Binding failed for port eaa99398-192b-4f83-a3fb-a0ac0164fa23, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 790.596726] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.607s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.599917] env[62521]: DEBUG nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Build of instance f6a27e36-1728-4056-99a6-ed2114df64ce was re-scheduled: Binding failed for port eaa99398-192b-4f83-a3fb-a0ac0164fa23, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 790.600354] env[62521]: DEBUG nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 790.600587] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Acquiring lock "refresh_cache-f6a27e36-1728-4056-99a6-ed2114df64ce" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.600735] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Acquired lock "refresh_cache-f6a27e36-1728-4056-99a6-ed2114df64ce" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.600891] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.618869] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.619148] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.619312] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.619493] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.619637] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.619781] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.619983] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.620167] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.620336] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.620498] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.620667] env[62521]: DEBUG nova.virt.hardware [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.622965] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801e1b66-066e-47b4-85b8-dba0aac06521 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.635529] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a1224c-e30d-410a-abf1-aac590db9f62 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.650468] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.656289] env[62521]: DEBUG oslo.service.loopingcall [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.660101] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 790.660101] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08897400-e97a-4fff-a72e-a08bd9174b99 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.674562] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.674562] env[62521]: value = "task-1318444" [ 790.674562] env[62521]: _type = "Task" [ 790.674562] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.682701] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318444, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.784899] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.927081] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.948855] env[62521]: INFO nova.compute.manager [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] [instance: 63abbe0a-af36-4679-8cf1-87f9e9abf7b9] Took 1.03 seconds to deallocate network for instance. [ 791.134121] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.140936] env[62521]: DEBUG nova.compute.manager [req-b69b31b2-8378-4556-bb13-6b464e3e1ea7 req-6cc7072c-d1d8-474b-90b2-1eee882325a9 service nova] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Received event network-vif-deleted-3d591a87-7f16-445c-80f4-c1f04dafd7b6 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.183856] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318444, 'name': CreateVM_Task, 'duration_secs': 0.408758} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.186276] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 791.186873] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.187046] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.187417] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.187679] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1793a956-b017-4955-9690-49d9bafbc277 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.192163] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 791.192163] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a0e22b-61b4-62a4-acff-f111f9d042cd" [ 791.192163] env[62521]: _type = "Task" [ 791.192163] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.202294] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a0e22b-61b4-62a4-acff-f111f9d042cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.352605] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888b012b-3e48-4638-b35d-a325068e1168 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.362353] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ff8aef-f30d-4b39-9534-72687283ed85 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.389440] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021f6de2-0315-4925-9fca-5996a1bcf1b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.396722] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df312a9-33ad-4706-b602-6ff2668a3e71 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.410412] env[62521]: DEBUG nova.compute.provider_tree [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.412345] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.432663] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Releasing lock "refresh_cache-3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.433104] env[62521]: DEBUG nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 791.433324] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 791.433644] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e3a56fb-caae-44dc-b11f-0d74f42866ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.442868] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755acdc6-4d3d-4427-91d9-576e6103fc94 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.473088] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e could not be found. [ 791.473311] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.473491] env[62521]: INFO nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 791.473812] env[62521]: DEBUG oslo.service.loopingcall [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.474209] env[62521]: DEBUG nova.compute.manager [-] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.474328] env[62521]: DEBUG nova.network.neutron [-] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.502013] env[62521]: DEBUG nova.network.neutron [-] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.703675] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a0e22b-61b4-62a4-acff-f111f9d042cd, 'name': SearchDatastore_Task, 'duration_secs': 0.017089} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.705566] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.705859] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.709082] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.709082] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.709082] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.709082] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e32613dc-cf1a-487d-9985-61944820b5d2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.718031] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.718031] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 791.718031] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abac61f7-bd17-4390-98f6-6e9d981b7725 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.723058] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 791.723058] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521e4d1b-75c6-a215-13b5-d263eee93238" [ 791.723058] env[62521]: _type = "Task" [ 791.723058] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.732476] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521e4d1b-75c6-a215-13b5-d263eee93238, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.917529] env[62521]: DEBUG nova.scheduler.client.report [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.921117] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Releasing lock "refresh_cache-f6a27e36-1728-4056-99a6-ed2114df64ce" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.921390] env[62521]: DEBUG nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 791.921627] env[62521]: DEBUG nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.921872] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.939634] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.986221] env[62521]: INFO nova.scheduler.client.report [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Deleted allocations for instance 63abbe0a-af36-4679-8cf1-87f9e9abf7b9 [ 792.004098] env[62521]: DEBUG nova.network.neutron [-] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.234327] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521e4d1b-75c6-a215-13b5-d263eee93238, 'name': SearchDatastore_Task, 'duration_secs': 0.008564} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.235154] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55fe089c-43e5-470d-89ad-3abf46ce798c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.241067] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 792.241067] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525b4219-bee7-ea10-9375-3b11f2b4eb4e" [ 792.241067] env[62521]: _type = "Task" [ 792.241067] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.249324] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525b4219-bee7-ea10-9375-3b11f2b4eb4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.424639] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.828s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.425288] env[62521]: ERROR nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c0bc58a3-690e-4f33-8b5a-2c55766167f5, please check neutron logs for more information. [ 792.425288] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Traceback (most recent call last): [ 792.425288] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 792.425288] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] self.driver.spawn(context, instance, image_meta, [ 792.425288] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 792.425288] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 792.425288] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 792.425288] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] vm_ref = self.build_virtual_machine(instance, [ 792.425288] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 792.425288] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] vif_infos = vmwarevif.get_vif_info(self._session, [ 792.425288] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] for vif in network_info: [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] return self._sync_wrapper(fn, *args, **kwargs) [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] self.wait() [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] self[:] = self._gt.wait() [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] return self._exit_event.wait() [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] result = hub.switch() [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 792.425634] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] return self.greenlet.switch() [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] result = function(*args, **kwargs) [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] return func(*args, **kwargs) [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] raise e [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] nwinfo = self.network_api.allocate_for_instance( [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] created_port_ids = self._update_ports_for_instance( [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] with excutils.save_and_reraise_exception(): [ 792.425987] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] self.force_reraise() [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] raise self.value [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] updated_port = self._update_port( [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] _ensure_no_port_binding_failure(port) [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] raise exception.PortBindingFailed(port_id=port['id']) [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] nova.exception.PortBindingFailed: Binding failed for port c0bc58a3-690e-4f33-8b5a-2c55766167f5, please check neutron logs for more information. [ 792.426376] env[62521]: ERROR nova.compute.manager [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] [ 792.426872] env[62521]: DEBUG nova.compute.utils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Binding failed for port c0bc58a3-690e-4f33-8b5a-2c55766167f5, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 792.427362] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.525s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.430196] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Build of instance 6be7d826-e19b-431e-b714-25e646e6b7ca was re-scheduled: Binding failed for port c0bc58a3-690e-4f33-8b5a-2c55766167f5, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 792.430628] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 792.430853] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "refresh_cache-6be7d826-e19b-431e-b714-25e646e6b7ca" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.431008] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquired lock "refresh_cache-6be7d826-e19b-431e-b714-25e646e6b7ca" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.431179] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.441542] env[62521]: DEBUG nova.network.neutron [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.496429] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2a663ddb-e77b-4f03-949e-459d662ad735 tempest-ImagesOneServerTestJSON-701178316 tempest-ImagesOneServerTestJSON-701178316-project-member] Lock "63abbe0a-af36-4679-8cf1-87f9e9abf7b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 169.487s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.506788] env[62521]: INFO nova.compute.manager [-] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Took 1.03 seconds to deallocate network for instance. [ 792.509619] env[62521]: DEBUG nova.compute.claims [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Aborting claim: {{(pid=62521) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 792.509849] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.752767] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525b4219-bee7-ea10-9375-3b11f2b4eb4e, 'name': SearchDatastore_Task, 'duration_secs': 0.011425} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.752767] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.753103] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 868c7798-8d0f-4c31-88dc-5007454796f1/868c7798-8d0f-4c31-88dc-5007454796f1.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 792.753225] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2eaff682-b621-4d31-b83b-28fe8c4ea543 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.761774] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 792.761774] env[62521]: value = "task-1318445" [ 792.761774] env[62521]: _type = "Task" [ 792.761774] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.772669] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318445, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.944909] env[62521]: INFO nova.compute.manager [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] [instance: f6a27e36-1728-4056-99a6-ed2114df64ce] Took 1.02 seconds to deallocate network for instance. [ 792.969165] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.001584] env[62521]: DEBUG nova.compute.manager [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 793.105128] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.226747] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1eb4793-edd9-4415-99aa-cbefb17d0262 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.234665] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0063e254-334c-47e4-8748-94f989004a1c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.268381] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e7aff1-a16a-4e3c-996b-27c86329a0bf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.275566] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318445, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485153} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.277898] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 868c7798-8d0f-4c31-88dc-5007454796f1/868c7798-8d0f-4c31-88dc-5007454796f1.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 793.279200] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.279499] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a4ff83ea-7028-46cc-8740-40077a0059a6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.282083] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e34da90-e94d-44c6-b511-67c69c879f0a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.294955] env[62521]: DEBUG nova.compute.provider_tree [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.297959] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 793.297959] env[62521]: value = "task-1318446" [ 793.297959] env[62521]: _type = "Task" [ 793.297959] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.305682] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318446, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.409031] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.409031] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.522424] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.608840] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Releasing lock "refresh_cache-6be7d826-e19b-431e-b714-25e646e6b7ca" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.609111] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 793.609308] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.609477] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 793.626101] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.799516] env[62521]: DEBUG nova.scheduler.client.report [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.812110] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318446, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063673} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.812512] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.813432] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000fc40d-b5b9-4d2d-b54a-698c15993acd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.835375] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 868c7798-8d0f-4c31-88dc-5007454796f1/868c7798-8d0f-4c31-88dc-5007454796f1.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.838013] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf2ec201-6840-4a0c-b95b-713869ba64f8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.855505] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 793.855505] env[62521]: value = "task-1318447" [ 793.855505] env[62521]: _type = "Task" [ 793.855505] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.864386] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318447, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.984638] env[62521]: INFO nova.scheduler.client.report [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Deleted allocations for instance f6a27e36-1728-4056-99a6-ed2114df64ce [ 794.128562] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.309074] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.880s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.309074] env[62521]: ERROR nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bf459636-3719-4f72-9f4b-e2e9b27811dd, please check neutron logs for more information. [ 794.309074] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Traceback (most recent call last): [ 794.309074] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 794.309074] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] self.driver.spawn(context, instance, image_meta, [ 794.309074] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 794.309074] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 794.309074] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 794.309074] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] vm_ref = self.build_virtual_machine(instance, [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] vif_infos = vmwarevif.get_vif_info(self._session, [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] for vif in network_info: [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] return self._sync_wrapper(fn, *args, **kwargs) [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] self.wait() [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] self[:] = self._gt.wait() [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] return self._exit_event.wait() [ 794.309497] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] current.throw(*self._exc) [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] result = function(*args, **kwargs) [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] return func(*args, **kwargs) [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] raise e [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] nwinfo = self.network_api.allocate_for_instance( [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] created_port_ids = self._update_ports_for_instance( [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 794.309915] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] with excutils.save_and_reraise_exception(): [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] self.force_reraise() [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] raise self.value [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] updated_port = self._update_port( [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] _ensure_no_port_binding_failure(port) [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] raise exception.PortBindingFailed(port_id=port['id']) [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] nova.exception.PortBindingFailed: Binding failed for port bf459636-3719-4f72-9f4b-e2e9b27811dd, please check neutron logs for more information. [ 794.310614] env[62521]: ERROR nova.compute.manager [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] [ 794.311091] env[62521]: DEBUG nova.compute.utils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Binding failed for port bf459636-3719-4f72-9f4b-e2e9b27811dd, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 794.311710] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.865s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.314161] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Build of instance 590b2441-e627-463b-9722-d0fcb8fcd83a was re-scheduled: Binding failed for port bf459636-3719-4f72-9f4b-e2e9b27811dd, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 794.317060] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 794.317060] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquiring lock "refresh_cache-590b2441-e627-463b-9722-d0fcb8fcd83a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.317060] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Acquired lock "refresh_cache-590b2441-e627-463b-9722-d0fcb8fcd83a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.317060] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.373115] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318447, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.503597] env[62521]: DEBUG oslo_concurrency.lockutils [None req-176fd654-8485-4b18-ac04-eb6e5505f4f8 tempest-FloatingIPsAssociationTestJSON-1689899208 tempest-FloatingIPsAssociationTestJSON-1689899208-project-member] Lock "f6a27e36-1728-4056-99a6-ed2114df64ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.157s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.634279] env[62521]: INFO nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 6be7d826-e19b-431e-b714-25e646e6b7ca] Took 1.02 seconds to deallocate network for instance. [ 794.845894] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.875096] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318447, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.946216] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.008341] env[62521]: DEBUG nova.compute.manager [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 795.374879] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318447, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.449377] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Releasing lock "refresh_cache-590b2441-e627-463b-9722-d0fcb8fcd83a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.449647] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 795.449832] env[62521]: DEBUG nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 795.450012] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 795.473114] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.527999] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.663748] env[62521]: INFO nova.scheduler.client.report [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Deleted allocations for instance 6be7d826-e19b-431e-b714-25e646e6b7ca [ 795.854891] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 6be7d826-e19b-431e-b714-25e646e6b7ca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.878156] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318447, 'name': ReconfigVM_Task, 'duration_secs': 1.714273} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.878432] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 868c7798-8d0f-4c31-88dc-5007454796f1/868c7798-8d0f-4c31-88dc-5007454796f1.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.879037] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-044496dd-bcd4-444e-bbcb-9725a2143515 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.886116] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 795.886116] env[62521]: value = "task-1318448" [ 795.886116] env[62521]: _type = "Task" [ 795.886116] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.894622] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318448, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.975485] env[62521]: DEBUG nova.network.neutron [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.176667] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "6be7d826-e19b-431e-b714-25e646e6b7ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.660s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.357766] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 590b2441-e627-463b-9722-d0fcb8fcd83a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.359843] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 062f311c-a783-41cf-8d83-b3a8d4df14f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.359843] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 868c7798-8d0f-4c31-88dc-5007454796f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.359843] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance de583c0b-d147-4f54-bc90-49bdd71c56aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.359843] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.397281] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318448, 'name': Rename_Task, 'duration_secs': 0.144064} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.397621] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 796.397810] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a68b072-d121-41a6-aa0c-03becbc8cdcb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.406122] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 796.406122] env[62521]: value = "task-1318449" [ 796.406122] env[62521]: _type = "Task" [ 796.406122] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.418098] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318449, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.477840] env[62521]: INFO nova.compute.manager [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] [instance: 590b2441-e627-463b-9722-d0fcb8fcd83a] Took 1.03 seconds to deallocate network for instance. [ 796.681284] env[62521]: DEBUG nova.compute.manager [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.866576] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 50aef23e-9429-4499-9fc1-c31b97e23029 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.918044] env[62521]: DEBUG oslo_vmware.api [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318449, 'name': PowerOnVM_Task, 'duration_secs': 0.420293} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.918044] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 796.918309] env[62521]: DEBUG nova.compute.manager [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.919358] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b9c9c0-2903-4aad-99ed-a4f7775ee2ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.371834] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 797.449301] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.558660] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.568168] env[62521]: INFO nova.scheduler.client.report [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Deleted allocations for instance 590b2441-e627-463b-9722-d0fcb8fcd83a [ 797.880631] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance e8cabcdc-d728-434d-a40d-fe751db086c3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.082247] env[62521]: DEBUG oslo_concurrency.lockutils [None req-704644e3-3ed8-4389-9d98-82ef3465fa13 tempest-MultipleCreateTestJSON-1403609722 tempest-MultipleCreateTestJSON-1403609722-project-member] Lock "590b2441-e627-463b-9722-d0fcb8fcd83a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.538s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.082247] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "868c7798-8d0f-4c31-88dc-5007454796f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.082247] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "868c7798-8d0f-4c31-88dc-5007454796f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.082247] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "868c7798-8d0f-4c31-88dc-5007454796f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.082385] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "868c7798-8d0f-4c31-88dc-5007454796f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.082385] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "868c7798-8d0f-4c31-88dc-5007454796f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.089337] env[62521]: INFO nova.compute.manager [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Terminating instance [ 798.092532] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "refresh_cache-868c7798-8d0f-4c31-88dc-5007454796f1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.092715] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired lock "refresh_cache-868c7798-8d0f-4c31-88dc-5007454796f1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.092895] env[62521]: DEBUG nova.network.neutron [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 798.384279] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance b3e33bab-636f-42df-ac6c-667464312c11 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.594100] env[62521]: DEBUG nova.compute.manager [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 798.623212] env[62521]: DEBUG nova.network.neutron [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.717015] env[62521]: DEBUG nova.network.neutron [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.888016] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance af91de53-2ef8-48d1-b8c1-7bd0809f4ae4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.125131] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.222812] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Releasing lock "refresh_cache-868c7798-8d0f-4c31-88dc-5007454796f1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.222812] env[62521]: DEBUG nova.compute.manager [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 799.222812] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 799.222812] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b87ebf-8ca7-4246-83ff-e956c1351ae1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.234164] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 799.236719] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b7ce854-6ede-4606-bb57-0afebf705bb4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.249055] env[62521]: DEBUG oslo_vmware.api [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 799.249055] env[62521]: value = "task-1318450" [ 799.249055] env[62521]: _type = "Task" [ 799.249055] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.256925] env[62521]: DEBUG oslo_vmware.api [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318450, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.391656] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance d5f6492d-ca1b-4e74-b792-b04d55c33660 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.755790] env[62521]: DEBUG oslo_vmware.api [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318450, 'name': PowerOffVM_Task, 'duration_secs': 0.124362} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.757330] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 799.757573] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 799.757835] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-844114b3-aff8-405a-a88b-c8a48e8d1f1f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.784376] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 799.784919] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 799.785222] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Deleting the datastore file [datastore1] 868c7798-8d0f-4c31-88dc-5007454796f1 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 799.787109] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32fa3e3a-dcb2-4fbf-8204-467d484128cb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.794977] env[62521]: DEBUG oslo_vmware.api [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 799.794977] env[62521]: value = "task-1318452" [ 799.794977] env[62521]: _type = "Task" [ 799.794977] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.805542] env[62521]: DEBUG oslo_vmware.api [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318452, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.894718] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 4dcd1c6c-8726-42db-997a-e78ba1293310 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.306938] env[62521]: DEBUG oslo_vmware.api [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318452, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090735} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.308456] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 800.308456] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 800.308456] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 800.308456] env[62521]: INFO nova.compute.manager [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Took 1.09 seconds to destroy the instance on the hypervisor. [ 800.308456] env[62521]: DEBUG oslo.service.loopingcall [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.308456] env[62521]: DEBUG nova.compute.manager [-] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 800.309293] env[62521]: DEBUG nova.network.neutron [-] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 800.327966] env[62521]: DEBUG nova.network.neutron [-] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.399580] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.622389] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.622389] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.832985] env[62521]: DEBUG nova.network.neutron [-] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.903596] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 4afe09db-6c01-444f-a127-6e1f97794544 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.335112] env[62521]: INFO nova.compute.manager [-] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Took 1.03 seconds to deallocate network for instance. [ 801.378828] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquiring lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.378927] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.418210] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 583b1a1b-1817-45fc-aa4f-fc9de4b4a243 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.841807] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.920399] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 92629ebf-eb50-4862-abe2-01e210e302ad has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.423997] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.424259] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 802.424402] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 802.451509] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 802.466750] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 802.466940] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 802.480912] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 802.499738] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 802.723482] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6550382b-071c-43ff-83eb-af5f92969e2e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.733453] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f158f4-f4ae-4633-bf2f-c25a1d80ac73 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.763683] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1239920f-8a4b-418e-92ef-d124aca45063 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.770919] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737d5444-81d4-4adb-aa51-501d67802010 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.784801] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.288850] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.793758] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 803.797114] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.483s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.797114] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.891s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.797114] env[62521]: INFO nova.compute.claims [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.801642] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 803.801959] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Cleaning up deleted instances {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 804.311565] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] There are 4 instances to clean {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 804.311894] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7653f469-5ce5-4d74-9a3f-55a6adb3a255] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 804.816781] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 20b01df0-4ced-4668-a786-046251a0c7ed] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 805.124192] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.124750] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.173444] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6af0040-daf9-4b31-9344-4c899f905ef5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.181993] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1251369-8137-4970-8876-af2c9d877e34 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.211335] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8a8b47-9d80-4631-8754-d8d5e5692bfd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.218542] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54012ea8-e32d-4be6-b91d-75b778443890 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.232581] env[62521]: DEBUG nova.compute.provider_tree [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.318699] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: e976d2d2-aa69-473b-9dfc-e616b0132754] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 805.736412] env[62521]: DEBUG nova.scheduler.client.report [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.821247] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 4f3a0a33-43c0-42fb-8660-198ff4a8529a] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 806.240950] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.241637] env[62521]: DEBUG nova.compute.manager [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 806.244026] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.317s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.324488] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 806.324733] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Cleaning up deleted instances with incomplete migration {{(pid=62521) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 806.751827] env[62521]: DEBUG nova.compute.utils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 806.753329] env[62521]: DEBUG nova.compute.manager [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 806.753498] env[62521]: DEBUG nova.network.neutron [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 806.792121] env[62521]: DEBUG nova.policy [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'abe9f1dc47a04c94b5bacdac3035dce1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dad7ae7edfd74245968c6e023c359547', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 806.827297] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 806.963961] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc10ae38-da5e-4d5d-b939-fb2b2af74136 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.971809] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9259bafc-b093-4b65-9ec7-0ecae997c928 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.002682] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9612e8e-360f-4c19-acef-bbdf2754fe8b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.009920] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237dfe86-243f-41ea-a9ec-678d409a7a51 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.022876] env[62521]: DEBUG nova.compute.provider_tree [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.162920] env[62521]: DEBUG nova.network.neutron [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Successfully created port: e43e3e90-9be9-4689-84a7-aa2baabf3271 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.257538] env[62521]: DEBUG nova.compute.manager [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 807.527061] env[62521]: DEBUG nova.scheduler.client.report [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.031395] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.787s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.032096] env[62521]: ERROR nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5e83db33-8f13-4a83-b35d-e9811b71dc4a, please check neutron logs for more information. [ 808.032096] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Traceback (most recent call last): [ 808.032096] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 808.032096] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] self.driver.spawn(context, instance, image_meta, [ 808.032096] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 808.032096] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 808.032096] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 808.032096] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] vm_ref = self.build_virtual_machine(instance, [ 808.032096] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 808.032096] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] vif_infos = vmwarevif.get_vif_info(self._session, [ 808.032096] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] for vif in network_info: [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] return self._sync_wrapper(fn, *args, **kwargs) [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] self.wait() [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] self[:] = self._gt.wait() [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] return self._exit_event.wait() [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] result = hub.switch() [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 808.032373] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] return self.greenlet.switch() [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] result = function(*args, **kwargs) [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] return func(*args, **kwargs) [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] raise e [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] nwinfo = self.network_api.allocate_for_instance( [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] created_port_ids = self._update_ports_for_instance( [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] with excutils.save_and_reraise_exception(): [ 808.032644] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] self.force_reraise() [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] raise self.value [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] updated_port = self._update_port( [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] _ensure_no_port_binding_failure(port) [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] raise exception.PortBindingFailed(port_id=port['id']) [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] nova.exception.PortBindingFailed: Binding failed for port 5e83db33-8f13-4a83-b35d-e9811b71dc4a, please check neutron logs for more information. [ 808.032926] env[62521]: ERROR nova.compute.manager [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] [ 808.033216] env[62521]: DEBUG nova.compute.utils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Binding failed for port 5e83db33-8f13-4a83-b35d-e9811b71dc4a, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 808.034326] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.695s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.035990] env[62521]: INFO nova.compute.claims [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.038922] env[62521]: DEBUG nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Build of instance de583c0b-d147-4f54-bc90-49bdd71c56aa was re-scheduled: Binding failed for port 5e83db33-8f13-4a83-b35d-e9811b71dc4a, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 808.039267] env[62521]: DEBUG nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 808.039727] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Acquiring lock "refresh_cache-de583c0b-d147-4f54-bc90-49bdd71c56aa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.039727] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Acquired lock "refresh_cache-de583c0b-d147-4f54-bc90-49bdd71c56aa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.040568] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.267537] env[62521]: DEBUG nova.compute.manager [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 808.292229] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 808.292480] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 808.292634] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 808.292811] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 808.292952] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 808.293107] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 808.293309] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 808.293464] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 808.293636] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 808.293792] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 808.293963] env[62521]: DEBUG nova.virt.hardware [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 808.294840] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def6c677-1db6-4fb2-bf7a-ff45eb44ffea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.303239] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b6181e-fe3f-46ad-82c9-5424749a98f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.561944] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.578629] env[62521]: DEBUG nova.compute.manager [req-db7fa7d0-6d41-49d9-bd56-179dd903fbd6 req-5bc2c187-49de-4a4c-b168-7616bb3a3f8d service nova] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Received event network-vif-plugged-e43e3e90-9be9-4689-84a7-aa2baabf3271 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.578836] env[62521]: DEBUG oslo_concurrency.lockutils [req-db7fa7d0-6d41-49d9-bd56-179dd903fbd6 req-5bc2c187-49de-4a4c-b168-7616bb3a3f8d service nova] Acquiring lock "50aef23e-9429-4499-9fc1-c31b97e23029-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.579049] env[62521]: DEBUG oslo_concurrency.lockutils [req-db7fa7d0-6d41-49d9-bd56-179dd903fbd6 req-5bc2c187-49de-4a4c-b168-7616bb3a3f8d service nova] Lock "50aef23e-9429-4499-9fc1-c31b97e23029-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.579213] env[62521]: DEBUG oslo_concurrency.lockutils [req-db7fa7d0-6d41-49d9-bd56-179dd903fbd6 req-5bc2c187-49de-4a4c-b168-7616bb3a3f8d service nova] Lock "50aef23e-9429-4499-9fc1-c31b97e23029-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.579370] env[62521]: DEBUG nova.compute.manager [req-db7fa7d0-6d41-49d9-bd56-179dd903fbd6 req-5bc2c187-49de-4a4c-b168-7616bb3a3f8d service nova] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] No waiting events found dispatching network-vif-plugged-e43e3e90-9be9-4689-84a7-aa2baabf3271 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 808.579525] env[62521]: WARNING nova.compute.manager [req-db7fa7d0-6d41-49d9-bd56-179dd903fbd6 req-5bc2c187-49de-4a4c-b168-7616bb3a3f8d service nova] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Received unexpected event network-vif-plugged-e43e3e90-9be9-4689-84a7-aa2baabf3271 for instance with vm_state building and task_state spawning. [ 808.658383] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.675298] env[62521]: DEBUG nova.network.neutron [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Successfully updated port: e43e3e90-9be9-4689-84a7-aa2baabf3271 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.160642] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Releasing lock "refresh_cache-de583c0b-d147-4f54-bc90-49bdd71c56aa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.160875] env[62521]: DEBUG nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 809.161069] env[62521]: DEBUG nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.161239] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.175990] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.180589] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquiring lock "refresh_cache-50aef23e-9429-4499-9fc1-c31b97e23029" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.180589] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquired lock "refresh_cache-50aef23e-9429-4499-9fc1-c31b97e23029" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.180589] env[62521]: DEBUG nova.network.neutron [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.259028] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2023a85d-cc28-4d98-8586-621f54dc5d18 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.266248] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9750a1f-cca8-4009-9b62-a5c8bbd2e6ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.297152] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7553239-1c2d-4f38-952d-7a4c3360e708 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.304681] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639584b6-bce0-432a-8fd3-78702ff16cc3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.317492] env[62521]: DEBUG nova.compute.provider_tree [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.681944] env[62521]: DEBUG nova.network.neutron [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.716708] env[62521]: DEBUG nova.network.neutron [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.820462] env[62521]: DEBUG nova.scheduler.client.report [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.889042] env[62521]: DEBUG nova.network.neutron [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Updating instance_info_cache with network_info: [{"id": "e43e3e90-9be9-4689-84a7-aa2baabf3271", "address": "fa:16:3e:66:34:1e", "network": {"id": "3234b61c-375d-4947-a872-7233b28a6bc3", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1159302766-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dad7ae7edfd74245968c6e023c359547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape43e3e90-9b", "ovs_interfaceid": "e43e3e90-9be9-4689-84a7-aa2baabf3271", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.187216] env[62521]: INFO nova.compute.manager [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] [instance: de583c0b-d147-4f54-bc90-49bdd71c56aa] Took 1.03 seconds to deallocate network for instance. [ 810.326749] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.327285] env[62521]: DEBUG nova.compute.manager [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 810.331854] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.092s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.333360] env[62521]: INFO nova.compute.claims [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.391803] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Releasing lock "refresh_cache-50aef23e-9429-4499-9fc1-c31b97e23029" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.392191] env[62521]: DEBUG nova.compute.manager [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Instance network_info: |[{"id": "e43e3e90-9be9-4689-84a7-aa2baabf3271", "address": "fa:16:3e:66:34:1e", "network": {"id": "3234b61c-375d-4947-a872-7233b28a6bc3", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1159302766-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dad7ae7edfd74245968c6e023c359547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape43e3e90-9b", "ovs_interfaceid": "e43e3e90-9be9-4689-84a7-aa2baabf3271", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 810.392903] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:34:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e43e3e90-9be9-4689-84a7-aa2baabf3271', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.400595] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Creating folder: Project (dad7ae7edfd74245968c6e023c359547). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.111698] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03f075a5-1096-4337-8189-7cd4960cf753 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.120050] env[62521]: DEBUG nova.compute.utils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 811.123991] env[62521]: DEBUG nova.compute.manager [req-2ec9b267-c2b9-47dc-bdf0-98492299e3c1 req-0ba01778-2512-425c-a088-b7d75177114b service nova] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Received event network-changed-e43e3e90-9be9-4689-84a7-aa2baabf3271 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.123991] env[62521]: DEBUG nova.compute.manager [req-2ec9b267-c2b9-47dc-bdf0-98492299e3c1 req-0ba01778-2512-425c-a088-b7d75177114b service nova] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Refreshing instance network info cache due to event network-changed-e43e3e90-9be9-4689-84a7-aa2baabf3271. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.123991] env[62521]: DEBUG oslo_concurrency.lockutils [req-2ec9b267-c2b9-47dc-bdf0-98492299e3c1 req-0ba01778-2512-425c-a088-b7d75177114b service nova] Acquiring lock "refresh_cache-50aef23e-9429-4499-9fc1-c31b97e23029" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.123991] env[62521]: DEBUG oslo_concurrency.lockutils [req-2ec9b267-c2b9-47dc-bdf0-98492299e3c1 req-0ba01778-2512-425c-a088-b7d75177114b service nova] Acquired lock "refresh_cache-50aef23e-9429-4499-9fc1-c31b97e23029" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.123991] env[62521]: DEBUG nova.network.neutron [req-2ec9b267-c2b9-47dc-bdf0-98492299e3c1 req-0ba01778-2512-425c-a088-b7d75177114b service nova] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Refreshing network info cache for port e43e3e90-9be9-4689-84a7-aa2baabf3271 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.125313] env[62521]: DEBUG nova.compute.manager [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 811.125468] env[62521]: DEBUG nova.network.neutron [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 811.137491] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Created folder: Project (dad7ae7edfd74245968c6e023c359547) in parent group-v282025. [ 811.137823] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Creating folder: Instances. Parent ref: group-v282055. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.138565] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-928fe352-016e-4706-8623-84a5ecc04086 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.152021] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Created folder: Instances in parent group-v282055. [ 811.152021] env[62521]: DEBUG oslo.service.loopingcall [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.152021] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.152021] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e9aa708-ad95-4e16-a2e5-79457236c3ef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.170304] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.170304] env[62521]: value = "task-1318455" [ 811.170304] env[62521]: _type = "Task" [ 811.170304] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.174587] env[62521]: DEBUG nova.policy [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a28a24310a04c6d861292815c825dac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f029ddf5cd8448caad5ec23775b6503', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 811.180849] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318455, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.455265] env[62521]: DEBUG nova.network.neutron [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Successfully created port: 06aa91ed-3b1d-4c27-a91f-6246e75add4e {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 811.642347] env[62521]: DEBUG nova.compute.manager [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 811.654393] env[62521]: INFO nova.scheduler.client.report [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Deleted allocations for instance de583c0b-d147-4f54-bc90-49bdd71c56aa [ 811.680259] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318455, 'name': CreateVM_Task, 'duration_secs': 0.319158} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.682631] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 811.713980] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.713980] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.714125] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 811.714299] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-186d9989-aa61-4f27-8bc6-e76b7401705c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.719353] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for the task: (returnval){ [ 811.719353] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5273433d-0020-30ad-1b1f-ea49212ede56" [ 811.719353] env[62521]: _type = "Task" [ 811.719353] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.729093] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5273433d-0020-30ad-1b1f-ea49212ede56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.894239] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14096d2e-8e1b-4b3e-b869-7ddc666ecf98 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.902923] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373e2908-1329-4b94-94bc-14494868737e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.934086] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80f47ab-4693-421c-b269-3117242c68ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.941419] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768afff6-4c43-479a-b2ea-93c1a2041f5e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.954794] env[62521]: DEBUG nova.compute.provider_tree [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.165992] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cd7b536-8bd3-4313-81d7-8e4990ea81d2 tempest-ServersTestFqdnHostnames-1063604696 tempest-ServersTestFqdnHostnames-1063604696-project-member] Lock "de583c0b-d147-4f54-bc90-49bdd71c56aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.017s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.192079] env[62521]: DEBUG nova.network.neutron [req-2ec9b267-c2b9-47dc-bdf0-98492299e3c1 req-0ba01778-2512-425c-a088-b7d75177114b service nova] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Updated VIF entry in instance network info cache for port e43e3e90-9be9-4689-84a7-aa2baabf3271. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 812.192414] env[62521]: DEBUG nova.network.neutron [req-2ec9b267-c2b9-47dc-bdf0-98492299e3c1 req-0ba01778-2512-425c-a088-b7d75177114b service nova] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Updating instance_info_cache with network_info: [{"id": "e43e3e90-9be9-4689-84a7-aa2baabf3271", "address": "fa:16:3e:66:34:1e", "network": {"id": "3234b61c-375d-4947-a872-7233b28a6bc3", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1159302766-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dad7ae7edfd74245968c6e023c359547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape43e3e90-9b", "ovs_interfaceid": "e43e3e90-9be9-4689-84a7-aa2baabf3271", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.233836] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5273433d-0020-30ad-1b1f-ea49212ede56, 'name': SearchDatastore_Task, 'duration_secs': 0.009812} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.234463] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.234697] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.234932] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.235114] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.235921] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.235921] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71cf9e86-9072-45d8-b478-d7aad94479f6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.243998] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.244203] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.244901] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86f2c9eb-9772-4936-9752-4dbb153894cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.250311] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for the task: (returnval){ [ 812.250311] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5234a45e-92ab-273c-0b39-b7ee58e17b7e" [ 812.250311] env[62521]: _type = "Task" [ 812.250311] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.257745] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5234a45e-92ab-273c-0b39-b7ee58e17b7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.459993] env[62521]: DEBUG nova.scheduler.client.report [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.654673] env[62521]: DEBUG nova.compute.manager [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 812.668864] env[62521]: DEBUG nova.compute.manager [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.679221] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 812.679485] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 812.679602] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 812.679784] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 812.679929] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 812.680085] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 812.680421] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 812.680494] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 812.680624] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 812.680756] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 812.680924] env[62521]: DEBUG nova.virt.hardware [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 812.681791] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121c8bb7-9ade-4baa-a824-75746079db1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.689852] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e7576f-312e-4de6-a028-8321df48af74 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.695404] env[62521]: DEBUG oslo_concurrency.lockutils [req-2ec9b267-c2b9-47dc-bdf0-98492299e3c1 req-0ba01778-2512-425c-a088-b7d75177114b service nova] Releasing lock "refresh_cache-50aef23e-9429-4499-9fc1-c31b97e23029" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.761697] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5234a45e-92ab-273c-0b39-b7ee58e17b7e, 'name': SearchDatastore_Task, 'duration_secs': 0.008384} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.762676] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28587503-ec0d-4e86-9e17-0ca91c1dbf72 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.767811] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for the task: (returnval){ [ 812.767811] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52712b24-16c0-0dbd-762b-55929ece9f66" [ 812.767811] env[62521]: _type = "Task" [ 812.767811] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.779239] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52712b24-16c0-0dbd-762b-55929ece9f66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.868492] env[62521]: DEBUG nova.compute.manager [req-21de18c0-049a-4f2d-a808-a36c11ac7b92 req-3cee268c-37de-46b3-bd66-e4ddf9cc27d2 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Received event network-vif-plugged-06aa91ed-3b1d-4c27-a91f-6246e75add4e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.868692] env[62521]: DEBUG oslo_concurrency.lockutils [req-21de18c0-049a-4f2d-a808-a36c11ac7b92 req-3cee268c-37de-46b3-bd66-e4ddf9cc27d2 service nova] Acquiring lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.868896] env[62521]: DEBUG oslo_concurrency.lockutils [req-21de18c0-049a-4f2d-a808-a36c11ac7b92 req-3cee268c-37de-46b3-bd66-e4ddf9cc27d2 service nova] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.869072] env[62521]: DEBUG oslo_concurrency.lockutils [req-21de18c0-049a-4f2d-a808-a36c11ac7b92 req-3cee268c-37de-46b3-bd66-e4ddf9cc27d2 service nova] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.869243] env[62521]: DEBUG nova.compute.manager [req-21de18c0-049a-4f2d-a808-a36c11ac7b92 req-3cee268c-37de-46b3-bd66-e4ddf9cc27d2 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] No waiting events found dispatching network-vif-plugged-06aa91ed-3b1d-4c27-a91f-6246e75add4e {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.869407] env[62521]: WARNING nova.compute.manager [req-21de18c0-049a-4f2d-a808-a36c11ac7b92 req-3cee268c-37de-46b3-bd66-e4ddf9cc27d2 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Received unexpected event network-vif-plugged-06aa91ed-3b1d-4c27-a91f-6246e75add4e for instance with vm_state building and task_state spawning. [ 812.966807] env[62521]: DEBUG nova.network.neutron [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Successfully updated port: 06aa91ed-3b1d-4c27-a91f-6246e75add4e {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.969026] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.970060] env[62521]: DEBUG nova.compute.manager [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.976659] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.818s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.980444] env[62521]: INFO nova.compute.claims [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.191980] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.279481] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52712b24-16c0-0dbd-762b-55929ece9f66, 'name': SearchDatastore_Task, 'duration_secs': 0.008833} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.279752] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.280367] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 50aef23e-9429-4499-9fc1-c31b97e23029/50aef23e-9429-4499-9fc1-c31b97e23029.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.280666] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fbfd12c8-6601-43ff-8516-03943b08529a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.287698] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for the task: (returnval){ [ 813.287698] env[62521]: value = "task-1318456" [ 813.287698] env[62521]: _type = "Task" [ 813.287698] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.296171] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318456, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.468223] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.468515] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquired lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.468664] env[62521]: DEBUG nova.network.neutron [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.481814] env[62521]: DEBUG nova.compute.utils [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.483524] env[62521]: DEBUG nova.compute.manager [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 813.483722] env[62521]: DEBUG nova.network.neutron [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 813.527693] env[62521]: DEBUG nova.policy [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9f6acb92872f4375837e9df4b8021607', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a207fb2d461340c4b2fc6e456596d719', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.797725] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318456, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490879} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.798212] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 50aef23e-9429-4499-9fc1-c31b97e23029/50aef23e-9429-4499-9fc1-c31b97e23029.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 813.798486] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.798757] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f645b85-d756-4993-a5dd-35a39c65ecc8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.806345] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for the task: (returnval){ [ 813.806345] env[62521]: value = "task-1318457" [ 813.806345] env[62521]: _type = "Task" [ 813.806345] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.812304] env[62521]: DEBUG nova.network.neutron [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Successfully created port: 53471722-7f78-46af-bc13-02c61f597490 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.817482] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318457, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.987814] env[62521]: DEBUG nova.compute.manager [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 814.001749] env[62521]: DEBUG nova.network.neutron [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.284755] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce169cff-3c0b-4c50-93a0-c7f4d3bc95c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.289746] env[62521]: DEBUG nova.network.neutron [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updating instance_info_cache with network_info: [{"id": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "address": "fa:16:3e:41:9c:b1", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06aa91ed-3b", "ovs_interfaceid": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.304047] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37358f9-9bef-4ce4-ae78-0d6806956fec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.343798] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318457, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062808} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.343798] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.345028] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e84260-17ea-448b-bbb9-9b028461066c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.347393] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9382c30d-66fd-4472-b857-d62bf1e0699f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.364807] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d942247-0514-4b5e-bd3c-5e55fbcffd64 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.381017] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 50aef23e-9429-4499-9fc1-c31b97e23029/50aef23e-9429-4499-9fc1-c31b97e23029.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.381017] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e051fde-cc51-4c9f-a0c4-01386dc26039 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.405013] env[62521]: DEBUG nova.compute.provider_tree [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.405013] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for the task: (returnval){ [ 814.405013] env[62521]: value = "task-1318458" [ 814.405013] env[62521]: _type = "Task" [ 814.405013] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.412191] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318458, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.495914] env[62521]: INFO nova.virt.block_device [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Booting with volume 683a2b16-e997-4b24-90c9-559ca2510a58 at /dev/sda [ 814.578027] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-699b3b96-397a-403b-9031-cc42b00add3c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.587107] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365b13ac-56cf-4575-9d18-890b8b3045f7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.611272] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c9b8c15-dd12-44ad-9087-e0a5b0ea6423 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.619245] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e637522e-9910-4552-937c-85cc3128860b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.642546] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b2a34a-532d-4750-b44f-aff6f5a36686 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.649487] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8263e8a6-0768-4820-bbcc-8ee18b5506a8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.662830] env[62521]: DEBUG nova.virt.block_device [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Updating existing volume attachment record: 7cb7b81d-f79b-41f4-9a2f-a74cda4bba83 {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 814.796263] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Releasing lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.796621] env[62521]: DEBUG nova.compute.manager [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Instance network_info: |[{"id": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "address": "fa:16:3e:41:9c:b1", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06aa91ed-3b", "ovs_interfaceid": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 814.797128] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:9c:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '06aa91ed-3b1d-4c27-a91f-6246e75add4e', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.804492] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Creating folder: Project (5f029ddf5cd8448caad5ec23775b6503). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.804797] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-97a911ef-f781-42e2-af29-e85b4f30d696 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.814937] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Created folder: Project (5f029ddf5cd8448caad5ec23775b6503) in parent group-v282025. [ 814.815123] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Creating folder: Instances. Parent ref: group-v282058. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.815339] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dbb9d047-4aec-4465-aebb-b032e5cd10ea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.823160] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Created folder: Instances in parent group-v282058. [ 814.823388] env[62521]: DEBUG oslo.service.loopingcall [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.823575] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 814.823761] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-999d0fde-c9d6-47e9-a030-503929c84ac1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.842450] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 814.842450] env[62521]: value = "task-1318461" [ 814.842450] env[62521]: _type = "Task" [ 814.842450] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.850058] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318461, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.901300] env[62521]: DEBUG nova.compute.manager [req-5da22a74-934c-44b6-9737-e0d28a293f00 req-cd29fba1-e275-457f-9ef5-f7bace16827f service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Received event network-changed-06aa91ed-3b1d-4c27-a91f-6246e75add4e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.901300] env[62521]: DEBUG nova.compute.manager [req-5da22a74-934c-44b6-9737-e0d28a293f00 req-cd29fba1-e275-457f-9ef5-f7bace16827f service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Refreshing instance network info cache due to event network-changed-06aa91ed-3b1d-4c27-a91f-6246e75add4e. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 814.901523] env[62521]: DEBUG oslo_concurrency.lockutils [req-5da22a74-934c-44b6-9737-e0d28a293f00 req-cd29fba1-e275-457f-9ef5-f7bace16827f service nova] Acquiring lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.901682] env[62521]: DEBUG oslo_concurrency.lockutils [req-5da22a74-934c-44b6-9737-e0d28a293f00 req-cd29fba1-e275-457f-9ef5-f7bace16827f service nova] Acquired lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.901826] env[62521]: DEBUG nova.network.neutron [req-5da22a74-934c-44b6-9737-e0d28a293f00 req-cd29fba1-e275-457f-9ef5-f7bace16827f service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Refreshing network info cache for port 06aa91ed-3b1d-4c27-a91f-6246e75add4e {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 814.910711] env[62521]: DEBUG nova.scheduler.client.report [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.923972] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318458, 'name': ReconfigVM_Task, 'duration_secs': 0.287723} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.924435] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 50aef23e-9429-4499-9fc1-c31b97e23029/50aef23e-9429-4499-9fc1-c31b97e23029.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 814.925692] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eeaddec9-5520-420c-b6e4-a913580d3442 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.935688] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for the task: (returnval){ [ 814.935688] env[62521]: value = "task-1318462" [ 814.935688] env[62521]: _type = "Task" [ 814.935688] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.945023] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318462, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.327894] env[62521]: DEBUG nova.network.neutron [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Successfully updated port: 53471722-7f78-46af-bc13-02c61f597490 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.352384] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318461, 'name': CreateVM_Task, 'duration_secs': 0.332238} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.352556] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 815.353251] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.353413] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.353765] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 815.354323] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-539635a3-57b5-4496-ba3d-a657487ffaec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.358941] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 815.358941] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52793dda-c04e-aa44-3fa9-c2f4daa56a85" [ 815.358941] env[62521]: _type = "Task" [ 815.358941] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.368232] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52793dda-c04e-aa44-3fa9-c2f4daa56a85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.416294] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.416816] env[62521]: DEBUG nova.compute.manager [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 815.419314] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.909s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.445160] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318462, 'name': Rename_Task, 'duration_secs': 0.142339} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.445272] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 815.445506] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7df04ebd-d429-4949-9703-9ee5732d6bd8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.452697] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for the task: (returnval){ [ 815.452697] env[62521]: value = "task-1318463" [ 815.452697] env[62521]: _type = "Task" [ 815.452697] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.462506] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318463, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.601947] env[62521]: DEBUG nova.network.neutron [req-5da22a74-934c-44b6-9737-e0d28a293f00 req-cd29fba1-e275-457f-9ef5-f7bace16827f service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updated VIF entry in instance network info cache for port 06aa91ed-3b1d-4c27-a91f-6246e75add4e. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 815.602330] env[62521]: DEBUG nova.network.neutron [req-5da22a74-934c-44b6-9737-e0d28a293f00 req-cd29fba1-e275-457f-9ef5-f7bace16827f service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updating instance_info_cache with network_info: [{"id": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "address": "fa:16:3e:41:9c:b1", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06aa91ed-3b", "ovs_interfaceid": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.830960] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Acquiring lock "refresh_cache-e8cabcdc-d728-434d-a40d-fe751db086c3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.831448] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Acquired lock "refresh_cache-e8cabcdc-d728-434d-a40d-fe751db086c3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.831448] env[62521]: DEBUG nova.network.neutron [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.869877] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52793dda-c04e-aa44-3fa9-c2f4daa56a85, 'name': SearchDatastore_Task, 'duration_secs': 0.009648} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.870201] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.870430] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.870661] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.870820] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.870987] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.871252] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-523a7f39-33ca-4a80-9ca0-ac388f12f5fb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.888449] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.888646] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 815.889456] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57484607-dbb8-474b-8aa4-3b6e4464412e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.894753] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 815.894753] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f700a0-57fd-7a01-ac70-1f07da6590b5" [ 815.894753] env[62521]: _type = "Task" [ 815.894753] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.902479] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f700a0-57fd-7a01-ac70-1f07da6590b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.923602] env[62521]: DEBUG nova.compute.utils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.924745] env[62521]: DEBUG nova.compute.manager [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 815.924916] env[62521]: DEBUG nova.network.neutron [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 815.963194] env[62521]: DEBUG oslo_vmware.api [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318463, 'name': PowerOnVM_Task, 'duration_secs': 0.470588} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.963704] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 815.965129] env[62521]: INFO nova.compute.manager [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Took 7.70 seconds to spawn the instance on the hypervisor. [ 815.965129] env[62521]: DEBUG nova.compute.manager [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 815.965379] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ba4ace-ebcf-4b5c-a220-2046e99907c1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.970760] env[62521]: DEBUG nova.policy [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c63ec03cec5437285d6a99fc2a1d791', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '840dfca13260497b93990e4875cfd532', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 816.104469] env[62521]: DEBUG oslo_concurrency.lockutils [req-5da22a74-934c-44b6-9737-e0d28a293f00 req-cd29fba1-e275-457f-9ef5-f7bace16827f service nova] Releasing lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.222086] env[62521]: DEBUG nova.network.neutron [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Successfully created port: a2a1e281-79b9-4cf2-9229-c2e6acb45ec7 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.240615] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414c7835-2465-409d-87c9-3607f095a06d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.248730] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4865a040-8f83-4f10-ae5d-a12f337e1bc2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.305465] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0e1d2c-25aa-4a16-bd01-22f374c3d869 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.317031] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bf956a-b5cc-4398-b6fd-54e66f5875c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.328891] env[62521]: DEBUG nova.compute.provider_tree [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.406152] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f700a0-57fd-7a01-ac70-1f07da6590b5, 'name': SearchDatastore_Task, 'duration_secs': 0.023253} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.408075] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4965571a-5504-42b7-b54f-2b3009e42250 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.413470] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 816.413470] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5250bb14-ae51-072e-9ef5-0481f2463728" [ 816.413470] env[62521]: _type = "Task" [ 816.413470] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.421468] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5250bb14-ae51-072e-9ef5-0481f2463728, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.430133] env[62521]: DEBUG nova.compute.manager [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 816.490605] env[62521]: INFO nova.compute.manager [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Took 31.60 seconds to build instance. [ 816.543023] env[62521]: DEBUG nova.network.neutron [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.708044] env[62521]: DEBUG nova.network.neutron [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Updating instance_info_cache with network_info: [{"id": "53471722-7f78-46af-bc13-02c61f597490", "address": "fa:16:3e:8f:1a:f8", "network": {"id": "3d22562d-07d9-4b37-be52-90908a15d111", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-61847516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a207fb2d461340c4b2fc6e456596d719", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53471722-7f", "ovs_interfaceid": "53471722-7f78-46af-bc13-02c61f597490", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.811017] env[62521]: DEBUG nova.compute.manager [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.811192] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.811414] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.811571] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.811751] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.811894] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.812080] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.812298] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.812458] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.812627] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.812809] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.813013] env[62521]: DEBUG nova.virt.hardware [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.813915] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80460098-5e1e-451a-92fb-9178ef5942bb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.822021] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f9f40ff-d66c-4c7d-b4a6-f169832ffe57 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.835925] env[62521]: DEBUG nova.scheduler.client.report [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.924931] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5250bb14-ae51-072e-9ef5-0481f2463728, 'name': SearchDatastore_Task, 'duration_secs': 0.031373} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.925206] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.925455] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 538ed18e-920e-4ed4-a9e1-a6bfbae7f282/538ed18e-920e-4ed4-a9e1-a6bfbae7f282.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 816.925725] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5aa80b50-1c28-4f02-9e25-3c8a18ff26b1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.931986] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 816.931986] env[62521]: value = "task-1318464" [ 816.931986] env[62521]: _type = "Task" [ 816.931986] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.943820] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318464, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.992151] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5eb702ee-9c33-4ec7-b056-c9e2d6c63a6a tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lock "50aef23e-9429-4499-9fc1-c31b97e23029" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.677s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.042046] env[62521]: DEBUG nova.compute.manager [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Received event network-vif-plugged-53471722-7f78-46af-bc13-02c61f597490 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.042299] env[62521]: DEBUG oslo_concurrency.lockutils [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] Acquiring lock "e8cabcdc-d728-434d-a40d-fe751db086c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.042528] env[62521]: DEBUG oslo_concurrency.lockutils [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] Lock "e8cabcdc-d728-434d-a40d-fe751db086c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.042730] env[62521]: DEBUG oslo_concurrency.lockutils [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] Lock "e8cabcdc-d728-434d-a40d-fe751db086c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.042862] env[62521]: DEBUG nova.compute.manager [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] No waiting events found dispatching network-vif-plugged-53471722-7f78-46af-bc13-02c61f597490 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 817.043087] env[62521]: WARNING nova.compute.manager [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Received unexpected event network-vif-plugged-53471722-7f78-46af-bc13-02c61f597490 for instance with vm_state building and task_state spawning. [ 817.043350] env[62521]: DEBUG nova.compute.manager [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Received event network-changed-53471722-7f78-46af-bc13-02c61f597490 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.043427] env[62521]: DEBUG nova.compute.manager [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Refreshing instance network info cache due to event network-changed-53471722-7f78-46af-bc13-02c61f597490. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.043610] env[62521]: DEBUG oslo_concurrency.lockutils [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] Acquiring lock "refresh_cache-e8cabcdc-d728-434d-a40d-fe751db086c3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.210612] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Releasing lock "refresh_cache-e8cabcdc-d728-434d-a40d-fe751db086c3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.211011] env[62521]: DEBUG nova.compute.manager [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Instance network_info: |[{"id": "53471722-7f78-46af-bc13-02c61f597490", "address": "fa:16:3e:8f:1a:f8", "network": {"id": "3d22562d-07d9-4b37-be52-90908a15d111", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-61847516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a207fb2d461340c4b2fc6e456596d719", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53471722-7f", "ovs_interfaceid": "53471722-7f78-46af-bc13-02c61f597490", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 817.211338] env[62521]: DEBUG oslo_concurrency.lockutils [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] Acquired lock "refresh_cache-e8cabcdc-d728-434d-a40d-fe751db086c3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.211517] env[62521]: DEBUG nova.network.neutron [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Refreshing network info cache for port 53471722-7f78-46af-bc13-02c61f597490 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 817.212833] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:1a:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4576b9d4-535c-40aa-b078-246f671f216e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '53471722-7f78-46af-bc13-02c61f597490', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 817.227174] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Creating folder: Project (a207fb2d461340c4b2fc6e456596d719). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 817.231583] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc540e1c-39e9-47a5-b2db-38df7747b656 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.248256] env[62521]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 817.248445] env[62521]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62521) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 817.249154] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Folder already exists: Project (a207fb2d461340c4b2fc6e456596d719). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 817.249355] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Creating folder: Instances. Parent ref: group-v282046. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 817.249600] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-50170b77-39ff-46dc-b466-c4964bb40d69 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.260534] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Created folder: Instances in parent group-v282046. [ 817.260791] env[62521]: DEBUG oslo.service.loopingcall [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.260994] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 817.261248] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c40ecc2d-830e-4c26-aa4c-433a307c0763 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.285099] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 817.285099] env[62521]: value = "task-1318467" [ 817.285099] env[62521]: _type = "Task" [ 817.285099] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.293426] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318467, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.344232] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.344232] env[62521]: ERROR nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3d591a87-7f16-445c-80f4-c1f04dafd7b6, please check neutron logs for more information. [ 817.344232] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Traceback (most recent call last): [ 817.344232] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 817.344232] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] self.driver.spawn(context, instance, image_meta, [ 817.344232] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 817.344232] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 817.344232] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 817.344232] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] vm_ref = self.build_virtual_machine(instance, [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] vif_infos = vmwarevif.get_vif_info(self._session, [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] for vif in network_info: [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] return self._sync_wrapper(fn, *args, **kwargs) [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] self.wait() [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] self[:] = self._gt.wait() [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] return self._exit_event.wait() [ 817.344895] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] result = hub.switch() [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] return self.greenlet.switch() [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] result = function(*args, **kwargs) [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] return func(*args, **kwargs) [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] raise e [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] nwinfo = self.network_api.allocate_for_instance( [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 817.345247] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] created_port_ids = self._update_ports_for_instance( [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] with excutils.save_and_reraise_exception(): [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] self.force_reraise() [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] raise self.value [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] updated_port = self._update_port( [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] _ensure_no_port_binding_failure(port) [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 817.345616] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] raise exception.PortBindingFailed(port_id=port['id']) [ 817.346072] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] nova.exception.PortBindingFailed: Binding failed for port 3d591a87-7f16-445c-80f4-c1f04dafd7b6, please check neutron logs for more information. [ 817.346072] env[62521]: ERROR nova.compute.manager [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] [ 817.346072] env[62521]: DEBUG nova.compute.utils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Binding failed for port 3d591a87-7f16-445c-80f4-c1f04dafd7b6, please check neutron logs for more information. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 817.346793] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.824s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.348677] env[62521]: INFO nova.compute.claims [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.352821] env[62521]: DEBUG nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Build of instance 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e was re-scheduled: Binding failed for port 3d591a87-7f16-445c-80f4-c1f04dafd7b6, please check neutron logs for more information. {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 817.353476] env[62521]: DEBUG nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Unplugging VIFs for instance {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 817.353874] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Acquiring lock "refresh_cache-3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.354142] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Acquired lock "refresh_cache-3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.354497] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.441130] env[62521]: DEBUG nova.compute.manager [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.446592] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318464, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.477044] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.477314] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.477473] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.477654] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.477796] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.477954] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.478287] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.478456] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.478648] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.478784] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.479016] env[62521]: DEBUG nova.virt.hardware [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.479854] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d8eb79-4f77-4998-bcf6-36d8535bf61b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.489595] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2274463-d7ef-41b5-8da2-682f6b65f9d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.495139] env[62521]: DEBUG nova.compute.manager [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 817.554107] env[62521]: DEBUG nova.network.neutron [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Updated VIF entry in instance network info cache for port 53471722-7f78-46af-bc13-02c61f597490. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 817.554517] env[62521]: DEBUG nova.network.neutron [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Updating instance_info_cache with network_info: [{"id": "53471722-7f78-46af-bc13-02c61f597490", "address": "fa:16:3e:8f:1a:f8", "network": {"id": "3d22562d-07d9-4b37-be52-90908a15d111", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-61847516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a207fb2d461340c4b2fc6e456596d719", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53471722-7f", "ovs_interfaceid": "53471722-7f78-46af-bc13-02c61f597490", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.654210] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquiring lock "50aef23e-9429-4499-9fc1-c31b97e23029" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.654210] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lock "50aef23e-9429-4499-9fc1-c31b97e23029" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.654210] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquiring lock "50aef23e-9429-4499-9fc1-c31b97e23029-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.654413] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lock "50aef23e-9429-4499-9fc1-c31b97e23029-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.654480] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lock "50aef23e-9429-4499-9fc1-c31b97e23029-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.656921] env[62521]: INFO nova.compute.manager [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Terminating instance [ 817.660023] env[62521]: DEBUG nova.compute.manager [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 817.660023] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 817.660023] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e337f921-d2d7-42ca-9aa1-a68c7bb443af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.672075] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 817.672075] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c475e0a3-82cc-4d26-8a52-473df5b456db {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.679711] env[62521]: DEBUG oslo_vmware.api [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for the task: (returnval){ [ 817.679711] env[62521]: value = "task-1318468" [ 817.679711] env[62521]: _type = "Task" [ 817.679711] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.689369] env[62521]: DEBUG oslo_vmware.api [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318468, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.792357] env[62521]: DEBUG nova.compute.manager [req-dc132348-b865-4cba-9c19-ca2fea870799 req-648bc1a2-656b-463e-8c8e-8c1393915cc0 service nova] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Received event network-vif-plugged-a2a1e281-79b9-4cf2-9229-c2e6acb45ec7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.792580] env[62521]: DEBUG oslo_concurrency.lockutils [req-dc132348-b865-4cba-9c19-ca2fea870799 req-648bc1a2-656b-463e-8c8e-8c1393915cc0 service nova] Acquiring lock "b3e33bab-636f-42df-ac6c-667464312c11-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.792807] env[62521]: DEBUG oslo_concurrency.lockutils [req-dc132348-b865-4cba-9c19-ca2fea870799 req-648bc1a2-656b-463e-8c8e-8c1393915cc0 service nova] Lock "b3e33bab-636f-42df-ac6c-667464312c11-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.792973] env[62521]: DEBUG oslo_concurrency.lockutils [req-dc132348-b865-4cba-9c19-ca2fea870799 req-648bc1a2-656b-463e-8c8e-8c1393915cc0 service nova] Lock "b3e33bab-636f-42df-ac6c-667464312c11-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.793165] env[62521]: DEBUG nova.compute.manager [req-dc132348-b865-4cba-9c19-ca2fea870799 req-648bc1a2-656b-463e-8c8e-8c1393915cc0 service nova] [instance: b3e33bab-636f-42df-ac6c-667464312c11] No waiting events found dispatching network-vif-plugged-a2a1e281-79b9-4cf2-9229-c2e6acb45ec7 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 817.793389] env[62521]: WARNING nova.compute.manager [req-dc132348-b865-4cba-9c19-ca2fea870799 req-648bc1a2-656b-463e-8c8e-8c1393915cc0 service nova] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Received unexpected event network-vif-plugged-a2a1e281-79b9-4cf2-9229-c2e6acb45ec7 for instance with vm_state building and task_state spawning. [ 817.799534] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318467, 'name': CreateVM_Task, 'duration_secs': 0.515683} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.799685] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.800334] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'delete_on_termination': True, 'device_type': None, 'mount_device': '/dev/sda', 'guest_format': None, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282049', 'volume_id': '683a2b16-e997-4b24-90c9-559ca2510a58', 'name': 'volume-683a2b16-e997-4b24-90c9-559ca2510a58', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8cabcdc-d728-434d-a40d-fe751db086c3', 'attached_at': '', 'detached_at': '', 'volume_id': '683a2b16-e997-4b24-90c9-559ca2510a58', 'serial': '683a2b16-e997-4b24-90c9-559ca2510a58'}, 'attachment_id': '7cb7b81d-f79b-41f4-9a2f-a74cda4bba83', 'volume_type': None}], 'swap': None} {{(pid=62521) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 817.800536] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Root volume attach. Driver type: vmdk {{(pid=62521) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 817.801295] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30fbb33-a137-41c6-81ee-bdf10e7be461 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.808898] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c6686da-54a0-44e1-96f7-5cbc33e09f84 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.814781] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cfbacbb-cc9a-4579-abab-e80a36885b71 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.820549] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-0a6d5075-75bd-412d-b287-7da935d7f31d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.827047] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for the task: (returnval){ [ 817.827047] env[62521]: value = "task-1318469" [ 817.827047] env[62521]: _type = "Task" [ 817.827047] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.834396] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318469, 'name': RelocateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.875885] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.942997] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318464, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550722} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.942997] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 538ed18e-920e-4ed4-a9e1-a6bfbae7f282/538ed18e-920e-4ed4-a9e1-a6bfbae7f282.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 817.943253] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.943324] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf343f2c-ad33-485e-8f08-a84920fcaa74 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.949908] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 817.949908] env[62521]: value = "task-1318470" [ 817.949908] env[62521]: _type = "Task" [ 817.949908] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.950489] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.959814] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318470, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.030063] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.056617] env[62521]: DEBUG oslo_concurrency.lockutils [req-070c21ee-1dab-4b5f-9d42-f9c5809d6dd4 req-2efc504f-c862-412b-9967-c8351ad404dc service nova] Releasing lock "refresh_cache-e8cabcdc-d728-434d-a40d-fe751db086c3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.191224] env[62521]: DEBUG oslo_vmware.api [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318468, 'name': PowerOffVM_Task, 'duration_secs': 0.235773} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.191425] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 818.191602] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 818.192053] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8ff1592-5050-4ad4-bdca-97fe6209b0b1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.245451] env[62521]: DEBUG nova.network.neutron [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Successfully updated port: a2a1e281-79b9-4cf2-9229-c2e6acb45ec7 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 818.302371] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 818.302710] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 818.302920] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Deleting the datastore file [datastore2] 50aef23e-9429-4499-9fc1-c31b97e23029 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 818.303213] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f4d6386-a97f-4d7d-9abe-8861608ac757 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.309956] env[62521]: DEBUG oslo_vmware.api [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for the task: (returnval){ [ 818.309956] env[62521]: value = "task-1318472" [ 818.309956] env[62521]: _type = "Task" [ 818.309956] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.318964] env[62521]: DEBUG oslo_vmware.api [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318472, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.339314] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318469, 'name': RelocateVM_Task} progress is 42%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.455010] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Releasing lock "refresh_cache-3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.455390] env[62521]: DEBUG nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62521) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 818.455510] env[62521]: DEBUG nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 818.455713] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 818.466479] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318470, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069246} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.466776] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.467605] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cef3b02-91ca-4c48-9412-5ccdad9320c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.492858] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 538ed18e-920e-4ed4-a9e1-a6bfbae7f282/538ed18e-920e-4ed4-a9e1-a6bfbae7f282.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.496676] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.498377] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89d5a70d-7f8e-4117-a7c7-0e7c508486be {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.514671] env[62521]: DEBUG nova.network.neutron [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.523027] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 818.523027] env[62521]: value = "task-1318473" [ 818.523027] env[62521]: _type = "Task" [ 818.523027] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.536567] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318473, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.657375] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e4886f-1e75-4246-9ca5-70da09d64770 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.665614] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b067df-213e-47de-b162-b22373a96c94 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.696096] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75abe5f5-8ee2-4eba-b0e2-ec8a1a2b6eda {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.704447] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33c4fdc-9c99-420f-9728-94079b5abb8c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.720465] env[62521]: DEBUG nova.compute.provider_tree [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.748498] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "refresh_cache-b3e33bab-636f-42df-ac6c-667464312c11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.748797] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired lock "refresh_cache-b3e33bab-636f-42df-ac6c-667464312c11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.749088] env[62521]: DEBUG nova.network.neutron [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 818.821911] env[62521]: DEBUG oslo_vmware.api [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Task: {'id': task-1318472, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259757} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.822215] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 818.822399] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 818.822571] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 818.822744] env[62521]: INFO nova.compute.manager [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Took 1.16 seconds to destroy the instance on the hypervisor. [ 818.822986] env[62521]: DEBUG oslo.service.loopingcall [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.823199] env[62521]: DEBUG nova.compute.manager [-] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 818.823289] env[62521]: DEBUG nova.network.neutron [-] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 818.836823] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318469, 'name': RelocateVM_Task} progress is 54%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.017376] env[62521]: INFO nova.compute.manager [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] [instance: 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e] Took 0.56 seconds to deallocate network for instance. [ 819.032428] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318473, 'name': ReconfigVM_Task, 'duration_secs': 0.371443} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.032720] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 538ed18e-920e-4ed4-a9e1-a6bfbae7f282/538ed18e-920e-4ed4-a9e1-a6bfbae7f282.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.034032] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1eff6950-f099-4e7b-8c4d-23c4ec1e4439 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.041745] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 819.041745] env[62521]: value = "task-1318474" [ 819.041745] env[62521]: _type = "Task" [ 819.041745] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.053373] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318474, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.225028] env[62521]: DEBUG nova.scheduler.client.report [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.246770] env[62521]: DEBUG nova.compute.manager [req-05e5755e-ddeb-448e-ab2e-c222b74f0345 req-25876db3-eb61-400c-9c6f-abd677c5f30b service nova] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Received event network-changed-a2a1e281-79b9-4cf2-9229-c2e6acb45ec7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.247279] env[62521]: DEBUG nova.compute.manager [req-05e5755e-ddeb-448e-ab2e-c222b74f0345 req-25876db3-eb61-400c-9c6f-abd677c5f30b service nova] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Refreshing instance network info cache due to event network-changed-a2a1e281-79b9-4cf2-9229-c2e6acb45ec7. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.247500] env[62521]: DEBUG oslo_concurrency.lockutils [req-05e5755e-ddeb-448e-ab2e-c222b74f0345 req-25876db3-eb61-400c-9c6f-abd677c5f30b service nova] Acquiring lock "refresh_cache-b3e33bab-636f-42df-ac6c-667464312c11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.294212] env[62521]: DEBUG nova.network.neutron [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.338232] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318469, 'name': RelocateVM_Task} progress is 69%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.559483] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318474, 'name': Rename_Task, 'duration_secs': 0.178887} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.560631] env[62521]: DEBUG nova.network.neutron [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Updating instance_info_cache with network_info: [{"id": "a2a1e281-79b9-4cf2-9229-c2e6acb45ec7", "address": "fa:16:3e:f9:10:a1", "network": {"id": "5b74eda2-668c-4d8f-b4d8-a5e748613ee1", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-574755447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "840dfca13260497b93990e4875cfd532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2a1e281-79", "ovs_interfaceid": "a2a1e281-79b9-4cf2-9229-c2e6acb45ec7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.562029] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.562488] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4d0f393-bbf3-47a4-8504-563915a84f9f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.570602] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 819.570602] env[62521]: value = "task-1318475" [ 819.570602] env[62521]: _type = "Task" [ 819.570602] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.580927] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318475, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.731282] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.731931] env[62521]: DEBUG nova.compute.manager [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.735637] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.208s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.737067] env[62521]: INFO nova.compute.claims [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.785909] env[62521]: DEBUG nova.network.neutron [-] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.841268] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318469, 'name': RelocateVM_Task} progress is 82%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.051600] env[62521]: INFO nova.scheduler.client.report [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Deleted allocations for instance 3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e [ 820.063584] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Releasing lock "refresh_cache-b3e33bab-636f-42df-ac6c-667464312c11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.064444] env[62521]: DEBUG nova.compute.manager [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Instance network_info: |[{"id": "a2a1e281-79b9-4cf2-9229-c2e6acb45ec7", "address": "fa:16:3e:f9:10:a1", "network": {"id": "5b74eda2-668c-4d8f-b4d8-a5e748613ee1", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-574755447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "840dfca13260497b93990e4875cfd532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2a1e281-79", "ovs_interfaceid": "a2a1e281-79b9-4cf2-9229-c2e6acb45ec7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 820.064444] env[62521]: DEBUG oslo_concurrency.lockutils [req-05e5755e-ddeb-448e-ab2e-c222b74f0345 req-25876db3-eb61-400c-9c6f-abd677c5f30b service nova] Acquired lock "refresh_cache-b3e33bab-636f-42df-ac6c-667464312c11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.064578] env[62521]: DEBUG nova.network.neutron [req-05e5755e-ddeb-448e-ab2e-c222b74f0345 req-25876db3-eb61-400c-9c6f-abd677c5f30b service nova] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Refreshing network info cache for port a2a1e281-79b9-4cf2-9229-c2e6acb45ec7 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.065515] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:10:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '415e68b4-3766-4359-afe2-f8563910d98c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a2a1e281-79b9-4cf2-9229-c2e6acb45ec7', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.073016] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Creating folder: Project (840dfca13260497b93990e4875cfd532). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.074237] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a12cef13-7523-44ce-917f-e41d4b55070d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.086435] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318475, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.089840] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Created folder: Project (840dfca13260497b93990e4875cfd532) in parent group-v282025. [ 820.090075] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Creating folder: Instances. Parent ref: group-v282063. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.090600] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a098a52-7faf-4f4e-bacb-c0243b644456 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.100182] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Created folder: Instances in parent group-v282063. [ 820.100522] env[62521]: DEBUG oslo.service.loopingcall [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.100742] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 820.100976] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8959112-26b7-4ee2-abfa-ef92cc9253b1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.121540] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.121540] env[62521]: value = "task-1318478" [ 820.121540] env[62521]: _type = "Task" [ 820.121540] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.129933] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318478, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.241158] env[62521]: DEBUG nova.compute.utils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.245283] env[62521]: DEBUG nova.compute.manager [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.245551] env[62521]: DEBUG nova.network.neutron [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 820.288580] env[62521]: INFO nova.compute.manager [-] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Took 1.47 seconds to deallocate network for instance. [ 820.311239] env[62521]: DEBUG nova.policy [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba9b0b82d53e4502bef18a9f031fcc9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5780fc8ecfd64161bc2ea0b3a092b4d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.341061] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318469, 'name': RelocateVM_Task} progress is 97%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.561468] env[62521]: DEBUG oslo_concurrency.lockutils [None req-95656571-990f-4be5-a1c4-149ae0a5840c tempest-ServerRescueTestJSON-911011158 tempest-ServerRescueTestJSON-911011158-project-member] Lock "3c2aaa4f-2a47-4304-acfe-8c67c22f5a7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.237s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.587526] env[62521]: DEBUG oslo_vmware.api [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318475, 'name': PowerOnVM_Task, 'duration_secs': 0.588467} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.587881] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 820.588162] env[62521]: INFO nova.compute.manager [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Took 7.93 seconds to spawn the instance on the hypervisor. [ 820.588406] env[62521]: DEBUG nova.compute.manager [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.589230] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c179c15-f95f-4795-b3ed-a646c9bbc874 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.632095] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318478, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.745803] env[62521]: DEBUG nova.compute.manager [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.796125] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.801376] env[62521]: DEBUG nova.network.neutron [req-05e5755e-ddeb-448e-ab2e-c222b74f0345 req-25876db3-eb61-400c-9c6f-abd677c5f30b service nova] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Updated VIF entry in instance network info cache for port a2a1e281-79b9-4cf2-9229-c2e6acb45ec7. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 820.801376] env[62521]: DEBUG nova.network.neutron [req-05e5755e-ddeb-448e-ab2e-c222b74f0345 req-25876db3-eb61-400c-9c6f-abd677c5f30b service nova] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Updating instance_info_cache with network_info: [{"id": "a2a1e281-79b9-4cf2-9229-c2e6acb45ec7", "address": "fa:16:3e:f9:10:a1", "network": {"id": "5b74eda2-668c-4d8f-b4d8-a5e748613ee1", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-574755447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "840dfca13260497b93990e4875cfd532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2a1e281-79", "ovs_interfaceid": "a2a1e281-79b9-4cf2-9229-c2e6acb45ec7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.832138] env[62521]: DEBUG nova.network.neutron [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Successfully created port: 6cf47748-b470-4bc0-8a01-23c0154e45be {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.843976] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318469, 'name': RelocateVM_Task} progress is 97%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.024137] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2956837-6da1-413a-aa6a-ddd61d59c0dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.033429] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e9f62f-dfd1-47eb-818e-ed66b8720cb4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.065550] env[62521]: DEBUG nova.compute.manager [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.069062] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d41b89-9d24-40ab-956a-a4c048a4fc32 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.077532] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6249096a-59f4-4f7d-9431-e44d44acdf46 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.093325] env[62521]: DEBUG nova.compute.provider_tree [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.106727] env[62521]: INFO nova.compute.manager [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Took 33.79 seconds to build instance. [ 821.137022] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318478, 'name': CreateVM_Task, 'duration_secs': 0.974038} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.137022] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.137022] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.137022] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.137022] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.137296] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aef10e26-efef-4acd-a5ab-4ba2d317d361 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.141406] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 821.141406] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5264ebcb-4f63-68be-a8f7-598d5db39be9" [ 821.141406] env[62521]: _type = "Task" [ 821.141406] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.150824] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5264ebcb-4f63-68be-a8f7-598d5db39be9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.276871] env[62521]: DEBUG nova.compute.manager [req-60f68431-0933-4930-8064-16e58adbd226 req-2d42dd28-100e-4fd0-838b-27a7e50a50a9 service nova] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Received event network-vif-deleted-e43e3e90-9be9-4689-84a7-aa2baabf3271 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.305035] env[62521]: DEBUG oslo_concurrency.lockutils [req-05e5755e-ddeb-448e-ab2e-c222b74f0345 req-25876db3-eb61-400c-9c6f-abd677c5f30b service nova] Releasing lock "refresh_cache-b3e33bab-636f-42df-ac6c-667464312c11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.341723] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318469, 'name': RelocateVM_Task} progress is 98%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.590656] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.595751] env[62521]: DEBUG nova.scheduler.client.report [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.609710] env[62521]: DEBUG oslo_concurrency.lockutils [None req-25774ada-08e2-43e7-9e29-b88b6954e1a6 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.114s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.652218] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5264ebcb-4f63-68be-a8f7-598d5db39be9, 'name': SearchDatastore_Task, 'duration_secs': 0.021685} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.652524] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.652750] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 821.652975] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.653134] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.653312] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 821.654141] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d4b319c-875d-4d13-bced-668ad5011074 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.661701] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 821.661953] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 821.662612] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8d98c3a-bb03-43cf-8517-49a077516279 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.667531] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 821.667531] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52452c3d-9e3e-1929-152f-9c990d6e3ef4" [ 821.667531] env[62521]: _type = "Task" [ 821.667531] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.675053] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52452c3d-9e3e-1929-152f-9c990d6e3ef4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.769098] env[62521]: DEBUG nova.compute.manager [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.799171] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.799421] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.799573] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.799931] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.799931] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.800051] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.801143] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.801143] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.801143] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.801143] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.801143] env[62521]: DEBUG nova.virt.hardware [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.801743] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7690421-0ed7-4c67-994c-682a6ccfc805 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.811427] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1145ea09-fa5d-4fcc-8df4-8406004f6072 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.840402] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318469, 'name': RelocateVM_Task, 'duration_secs': 3.677241} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.840640] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 821.840884] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282049', 'volume_id': '683a2b16-e997-4b24-90c9-559ca2510a58', 'name': 'volume-683a2b16-e997-4b24-90c9-559ca2510a58', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8cabcdc-d728-434d-a40d-fe751db086c3', 'attached_at': '', 'detached_at': '', 'volume_id': '683a2b16-e997-4b24-90c9-559ca2510a58', 'serial': '683a2b16-e997-4b24-90c9-559ca2510a58'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 821.842545] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17381dba-9409-4ff9-8264-77cee50e2dae {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.860077] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0097445-bcd9-4ae5-98b7-a437c6d31252 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.885465] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] volume-683a2b16-e997-4b24-90c9-559ca2510a58/volume-683a2b16-e997-4b24-90c9-559ca2510a58.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.885804] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1dfe653-1e74-463c-84c2-b6a34fcc2907 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.905917] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for the task: (returnval){ [ 821.905917] env[62521]: value = "task-1318479" [ 821.905917] env[62521]: _type = "Task" [ 821.905917] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.914587] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318479, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.100606] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.101195] env[62521]: DEBUG nova.compute.manager [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.103941] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.657s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.104170] env[62521]: DEBUG nova.objects.instance [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62521) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 822.115137] env[62521]: DEBUG nova.compute.manager [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 822.180763] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52452c3d-9e3e-1929-152f-9c990d6e3ef4, 'name': SearchDatastore_Task, 'duration_secs': 0.008532} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.181839] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fae00ac-39cb-4b0d-beb8-d16e4e875154 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.189039] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 822.189039] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a77a02-05fe-749d-6f33-20716f4a0a2d" [ 822.189039] env[62521]: _type = "Task" [ 822.189039] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.198841] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a77a02-05fe-749d-6f33-20716f4a0a2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.420825] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318479, 'name': ReconfigVM_Task, 'duration_secs': 0.291654} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.421201] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Reconfigured VM instance instance-00000038 to attach disk [datastore2] volume-683a2b16-e997-4b24-90c9-559ca2510a58/volume-683a2b16-e997-4b24-90c9-559ca2510a58.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.428014] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39cfbe30-11a5-46a3-930d-7c7224bc767f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.448740] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for the task: (returnval){ [ 822.448740] env[62521]: value = "task-1318480" [ 822.448740] env[62521]: _type = "Task" [ 822.448740] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.454934] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318480, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.608831] env[62521]: DEBUG nova.compute.utils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.614329] env[62521]: DEBUG nova.compute.manager [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.614511] env[62521]: DEBUG nova.network.neutron [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.641976] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.673298] env[62521]: DEBUG nova.policy [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '937840982b914e68849319714a16b1a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea161b1af2bb4b9f8cc39b16582f7013', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.699990] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a77a02-05fe-749d-6f33-20716f4a0a2d, 'name': SearchDatastore_Task, 'duration_secs': 0.00942} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.700513] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.700831] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] b3e33bab-636f-42df-ac6c-667464312c11/b3e33bab-636f-42df-ac6c-667464312c11.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 822.701265] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f60c818-ce55-4d65-95a3-13043c0ca67b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.710121] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 822.710121] env[62521]: value = "task-1318481" [ 822.710121] env[62521]: _type = "Task" [ 822.710121] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.722960] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318481, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.760313] env[62521]: DEBUG nova.compute.manager [req-000aab7e-ad8c-4131-829c-299eb2ab5262 req-37cabebb-5262-42dc-bc82-55a091392bb4 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Received event network-vif-plugged-6cf47748-b470-4bc0-8a01-23c0154e45be {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.760577] env[62521]: DEBUG oslo_concurrency.lockutils [req-000aab7e-ad8c-4131-829c-299eb2ab5262 req-37cabebb-5262-42dc-bc82-55a091392bb4 service nova] Acquiring lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.760839] env[62521]: DEBUG oslo_concurrency.lockutils [req-000aab7e-ad8c-4131-829c-299eb2ab5262 req-37cabebb-5262-42dc-bc82-55a091392bb4 service nova] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.760999] env[62521]: DEBUG oslo_concurrency.lockutils [req-000aab7e-ad8c-4131-829c-299eb2ab5262 req-37cabebb-5262-42dc-bc82-55a091392bb4 service nova] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.761191] env[62521]: DEBUG nova.compute.manager [req-000aab7e-ad8c-4131-829c-299eb2ab5262 req-37cabebb-5262-42dc-bc82-55a091392bb4 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] No waiting events found dispatching network-vif-plugged-6cf47748-b470-4bc0-8a01-23c0154e45be {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.761361] env[62521]: WARNING nova.compute.manager [req-000aab7e-ad8c-4131-829c-299eb2ab5262 req-37cabebb-5262-42dc-bc82-55a091392bb4 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Received unexpected event network-vif-plugged-6cf47748-b470-4bc0-8a01-23c0154e45be for instance with vm_state building and task_state spawning. [ 822.960498] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318480, 'name': ReconfigVM_Task, 'duration_secs': 0.197632} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.960846] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282049', 'volume_id': '683a2b16-e997-4b24-90c9-559ca2510a58', 'name': 'volume-683a2b16-e997-4b24-90c9-559ca2510a58', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8cabcdc-d728-434d-a40d-fe751db086c3', 'attached_at': '', 'detached_at': '', 'volume_id': '683a2b16-e997-4b24-90c9-559ca2510a58', 'serial': '683a2b16-e997-4b24-90c9-559ca2510a58'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 822.961725] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a75f6b7a-034a-4fea-a19a-06c23797796e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.971159] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for the task: (returnval){ [ 822.971159] env[62521]: value = "task-1318482" [ 822.971159] env[62521]: _type = "Task" [ 822.971159] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.985224] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318482, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.053224] env[62521]: DEBUG nova.network.neutron [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Successfully created port: 7e00f1c5-5c15-430f-baff-cf464bf5d9b3 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.117829] env[62521]: DEBUG nova.compute.manager [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 823.124231] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fa937461-40c8-4394-8d8e-8ee373cbd935 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.128919] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.570s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.130473] env[62521]: INFO nova.compute.claims [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.169505] env[62521]: DEBUG nova.compute.manager [req-353467d8-56af-4f6c-8a34-13919b44a1aa req-879543b6-e7d0-4856-ae56-b878019af0f6 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Received event network-changed-06aa91ed-3b1d-4c27-a91f-6246e75add4e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.169711] env[62521]: DEBUG nova.compute.manager [req-353467d8-56af-4f6c-8a34-13919b44a1aa req-879543b6-e7d0-4856-ae56-b878019af0f6 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Refreshing instance network info cache due to event network-changed-06aa91ed-3b1d-4c27-a91f-6246e75add4e. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.169931] env[62521]: DEBUG oslo_concurrency.lockutils [req-353467d8-56af-4f6c-8a34-13919b44a1aa req-879543b6-e7d0-4856-ae56-b878019af0f6 service nova] Acquiring lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.171728] env[62521]: DEBUG oslo_concurrency.lockutils [req-353467d8-56af-4f6c-8a34-13919b44a1aa req-879543b6-e7d0-4856-ae56-b878019af0f6 service nova] Acquired lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.171971] env[62521]: DEBUG nova.network.neutron [req-353467d8-56af-4f6c-8a34-13919b44a1aa req-879543b6-e7d0-4856-ae56-b878019af0f6 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Refreshing network info cache for port 06aa91ed-3b1d-4c27-a91f-6246e75add4e {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 823.222454] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318481, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.406761] env[62521]: DEBUG nova.network.neutron [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Successfully updated port: 6cf47748-b470-4bc0-8a01-23c0154e45be {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 823.482815] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318482, 'name': Rename_Task, 'duration_secs': 0.278418} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.483354] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.483709] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e56e215-a2e6-4246-b8ad-06700e41d80e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.492026] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for the task: (returnval){ [ 823.492026] env[62521]: value = "task-1318483" [ 823.492026] env[62521]: _type = "Task" [ 823.492026] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.498373] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318483, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.724144] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318481, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517683} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.724433] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] b3e33bab-636f-42df-ac6c-667464312c11/b3e33bab-636f-42df-ac6c-667464312c11.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 823.724647] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 823.724904] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00e093f4-eba1-45bd-a712-d562dbbbc928 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.732215] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 823.732215] env[62521]: value = "task-1318484" [ 823.732215] env[62521]: _type = "Task" [ 823.732215] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.743854] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318484, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.911572] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "refresh_cache-af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.911876] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "refresh_cache-af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.912392] env[62521]: DEBUG nova.network.neutron [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.000786] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318483, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.006882] env[62521]: DEBUG nova.network.neutron [req-353467d8-56af-4f6c-8a34-13919b44a1aa req-879543b6-e7d0-4856-ae56-b878019af0f6 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updated VIF entry in instance network info cache for port 06aa91ed-3b1d-4c27-a91f-6246e75add4e. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 824.007304] env[62521]: DEBUG nova.network.neutron [req-353467d8-56af-4f6c-8a34-13919b44a1aa req-879543b6-e7d0-4856-ae56-b878019af0f6 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updating instance_info_cache with network_info: [{"id": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "address": "fa:16:3e:41:9c:b1", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06aa91ed-3b", "ovs_interfaceid": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.132990] env[62521]: DEBUG nova.compute.manager [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.157560] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.157886] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.158020] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.158225] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.158436] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.158549] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.158776] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.158947] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.159151] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.159735] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.159735] env[62521]: DEBUG nova.virt.hardware [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.160361] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aebea2c-46fa-4ff1-9dad-fe0259b2e7a9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.170609] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521d2f7f-f36b-45b9-b4e8-eac51021191e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.246919] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318484, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071162} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.247354] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.248175] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3bf430-464c-47f6-9b8d-d07fac6c44d3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.274608] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] b3e33bab-636f-42df-ac6c-667464312c11/b3e33bab-636f-42df-ac6c-667464312c11.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.276557] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-867dfebc-76e4-438e-ad03-f139f3e93cd7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.298163] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 824.298163] env[62521]: value = "task-1318485" [ 824.298163] env[62521]: _type = "Task" [ 824.298163] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.307017] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318485, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.406258] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c83349-849a-4251-974e-548907301111 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.414295] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f1cda7-027b-4967-8092-db61187e5e0a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.445986] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6206647d-c289-48a0-ade5-791eb5472e84 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.453847] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b396325a-db4f-4e63-8831-94e859348d5f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.459271] env[62521]: DEBUG nova.network.neutron [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.473210] env[62521]: DEBUG nova.compute.provider_tree [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.501805] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318483, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.510510] env[62521]: DEBUG oslo_concurrency.lockutils [req-353467d8-56af-4f6c-8a34-13919b44a1aa req-879543b6-e7d0-4856-ae56-b878019af0f6 service nova] Releasing lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.568643] env[62521]: DEBUG nova.compute.manager [req-bc2114e6-6c40-461f-a0d1-51ecc9d120fe req-2101e5d1-15c3-4a74-9d0f-89faba4df6aa service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Received event network-vif-plugged-7e00f1c5-5c15-430f-baff-cf464bf5d9b3 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.568643] env[62521]: DEBUG oslo_concurrency.lockutils [req-bc2114e6-6c40-461f-a0d1-51ecc9d120fe req-2101e5d1-15c3-4a74-9d0f-89faba4df6aa service nova] Acquiring lock "d5f6492d-ca1b-4e74-b792-b04d55c33660-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.568643] env[62521]: DEBUG oslo_concurrency.lockutils [req-bc2114e6-6c40-461f-a0d1-51ecc9d120fe req-2101e5d1-15c3-4a74-9d0f-89faba4df6aa service nova] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.568643] env[62521]: DEBUG oslo_concurrency.lockutils [req-bc2114e6-6c40-461f-a0d1-51ecc9d120fe req-2101e5d1-15c3-4a74-9d0f-89faba4df6aa service nova] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.568643] env[62521]: DEBUG nova.compute.manager [req-bc2114e6-6c40-461f-a0d1-51ecc9d120fe req-2101e5d1-15c3-4a74-9d0f-89faba4df6aa service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] No waiting events found dispatching network-vif-plugged-7e00f1c5-5c15-430f-baff-cf464bf5d9b3 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 824.568987] env[62521]: WARNING nova.compute.manager [req-bc2114e6-6c40-461f-a0d1-51ecc9d120fe req-2101e5d1-15c3-4a74-9d0f-89faba4df6aa service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Received unexpected event network-vif-plugged-7e00f1c5-5c15-430f-baff-cf464bf5d9b3 for instance with vm_state building and task_state spawning. [ 824.627541] env[62521]: DEBUG nova.network.neutron [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Successfully updated port: 7e00f1c5-5c15-430f-baff-cf464bf5d9b3 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 824.647027] env[62521]: DEBUG nova.network.neutron [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Updating instance_info_cache with network_info: [{"id": "6cf47748-b470-4bc0-8a01-23c0154e45be", "address": "fa:16:3e:4d:ae:61", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cf47748-b4", "ovs_interfaceid": "6cf47748-b470-4bc0-8a01-23c0154e45be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.812567] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.977190] env[62521]: DEBUG nova.scheduler.client.report [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.005063] env[62521]: DEBUG oslo_vmware.api [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318483, 'name': PowerOnVM_Task, 'duration_secs': 1.133063} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.005063] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.005063] env[62521]: INFO nova.compute.manager [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Took 8.19 seconds to spawn the instance on the hypervisor. [ 825.005063] env[62521]: DEBUG nova.compute.manager [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.005063] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe422e2-07d1-4da6-9403-65be43c2045b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.129985] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.130142] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.130301] env[62521]: DEBUG nova.network.neutron [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.149981] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "refresh_cache-af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.151041] env[62521]: DEBUG nova.compute.manager [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Instance network_info: |[{"id": "6cf47748-b470-4bc0-8a01-23c0154e45be", "address": "fa:16:3e:4d:ae:61", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cf47748-b4", "ovs_interfaceid": "6cf47748-b470-4bc0-8a01-23c0154e45be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 825.151271] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:ae:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd998416-f3d6-4a62-b828-5011063ce76a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6cf47748-b470-4bc0-8a01-23c0154e45be', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.158264] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Creating folder: Project (5780fc8ecfd64161bc2ea0b3a092b4d7). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.158777] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b6e73a4-4cdd-48f3-a059-107854e40f2b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.167855] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Created folder: Project (5780fc8ecfd64161bc2ea0b3a092b4d7) in parent group-v282025. [ 825.168047] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Creating folder: Instances. Parent ref: group-v282066. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.168281] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-affce3a8-32b8-4858-a5c3-fd311ba62b00 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.177751] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Created folder: Instances in parent group-v282066. [ 825.177984] env[62521]: DEBUG oslo.service.loopingcall [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.178182] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.178375] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64b0a0cb-257d-4b29-a928-249f11b8df6f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.197951] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.197951] env[62521]: value = "task-1318488" [ 825.197951] env[62521]: _type = "Task" [ 825.197951] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.205592] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318488, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.310067] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318485, 'name': ReconfigVM_Task, 'duration_secs': 0.851528} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.310434] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Reconfigured VM instance instance-00000039 to attach disk [datastore2] b3e33bab-636f-42df-ac6c-667464312c11/b3e33bab-636f-42df-ac6c-667464312c11.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 825.311148] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b100293e-d91c-484b-8c00-ef165b3bf6f5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.318027] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 825.318027] env[62521]: value = "task-1318489" [ 825.318027] env[62521]: _type = "Task" [ 825.318027] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.326255] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318489, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.379689] env[62521]: DEBUG nova.compute.manager [req-1c08bda8-fbb5-44d0-ad65-e0e9ff87b8e9 req-a1e1bb04-0694-4339-b3d2-12f0dffb56a7 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Received event network-changed-6cf47748-b470-4bc0-8a01-23c0154e45be {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.379965] env[62521]: DEBUG nova.compute.manager [req-1c08bda8-fbb5-44d0-ad65-e0e9ff87b8e9 req-a1e1bb04-0694-4339-b3d2-12f0dffb56a7 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Refreshing instance network info cache due to event network-changed-6cf47748-b470-4bc0-8a01-23c0154e45be. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 825.380235] env[62521]: DEBUG oslo_concurrency.lockutils [req-1c08bda8-fbb5-44d0-ad65-e0e9ff87b8e9 req-a1e1bb04-0694-4339-b3d2-12f0dffb56a7 service nova] Acquiring lock "refresh_cache-af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.380435] env[62521]: DEBUG oslo_concurrency.lockutils [req-1c08bda8-fbb5-44d0-ad65-e0e9ff87b8e9 req-a1e1bb04-0694-4339-b3d2-12f0dffb56a7 service nova] Acquired lock "refresh_cache-af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.380634] env[62521]: DEBUG nova.network.neutron [req-1c08bda8-fbb5-44d0-ad65-e0e9ff87b8e9 req-a1e1bb04-0694-4339-b3d2-12f0dffb56a7 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Refreshing network info cache for port 6cf47748-b470-4bc0-8a01-23c0154e45be {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.486023] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.486023] env[62521]: DEBUG nova.compute.manager [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.489033] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.364s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.498264] env[62521]: INFO nova.compute.claims [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.527409] env[62521]: INFO nova.compute.manager [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Took 37.30 seconds to build instance. [ 825.670214] env[62521]: DEBUG nova.network.neutron [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.710374] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318488, 'name': CreateVM_Task, 'duration_secs': 0.47438} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.710560] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.711431] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.711600] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.711956] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.712349] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-678c0f4c-ac01-4ea0-9a8f-b12db0e1850f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.718903] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 825.718903] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520a1d0a-5e59-3045-ee08-7661565e5bfe" [ 825.718903] env[62521]: _type = "Task" [ 825.718903] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.727134] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520a1d0a-5e59-3045-ee08-7661565e5bfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.829666] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318489, 'name': Rename_Task, 'duration_secs': 0.181185} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.829964] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 825.830262] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-583729e1-408c-4607-b7fd-728aab7c54a4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.836792] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 825.836792] env[62521]: value = "task-1318490" [ 825.836792] env[62521]: _type = "Task" [ 825.836792] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.846999] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.927458] env[62521]: DEBUG nova.network.neutron [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance_info_cache with network_info: [{"id": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "address": "fa:16:3e:d7:52:1f", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e00f1c5-5c", "ovs_interfaceid": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.037578] env[62521]: DEBUG nova.compute.utils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.041886] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a0b492ae-d695-4a7f-b484-097b9ad62cfe tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "e8cabcdc-d728-434d-a40d-fe751db086c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.512s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.042726] env[62521]: DEBUG nova.compute.manager [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.043053] env[62521]: DEBUG nova.network.neutron [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 826.175687] env[62521]: DEBUG nova.policy [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0e99fc3ead94078b185bc599ff247ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25420ed4f82e478989a011d40f037a24', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.198781] env[62521]: DEBUG nova.network.neutron [req-1c08bda8-fbb5-44d0-ad65-e0e9ff87b8e9 req-a1e1bb04-0694-4339-b3d2-12f0dffb56a7 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Updated VIF entry in instance network info cache for port 6cf47748-b470-4bc0-8a01-23c0154e45be. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 826.199203] env[62521]: DEBUG nova.network.neutron [req-1c08bda8-fbb5-44d0-ad65-e0e9ff87b8e9 req-a1e1bb04-0694-4339-b3d2-12f0dffb56a7 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Updating instance_info_cache with network_info: [{"id": "6cf47748-b470-4bc0-8a01-23c0154e45be", "address": "fa:16:3e:4d:ae:61", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cf47748-b4", "ovs_interfaceid": "6cf47748-b470-4bc0-8a01-23c0154e45be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.230995] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520a1d0a-5e59-3045-ee08-7661565e5bfe, 'name': SearchDatastore_Task, 'duration_secs': 0.024406} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.230995] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.230995] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.230995] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.231302] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.231302] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.231302] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-500d9dee-1eff-4e3d-b305-747567c1faf1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.242403] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.242590] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.243320] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f1bb29f-24d0-41f9-b5d3-3a3ab14276b5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.249186] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 826.249186] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52445020-80d1-8fdb-3db4-9281f48766c5" [ 826.249186] env[62521]: _type = "Task" [ 826.249186] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.258051] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52445020-80d1-8fdb-3db4-9281f48766c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.346658] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318490, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.434059] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.434059] env[62521]: DEBUG nova.compute.manager [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Instance network_info: |[{"id": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "address": "fa:16:3e:d7:52:1f", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e00f1c5-5c", "ovs_interfaceid": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 826.434282] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:52:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc7aa55d-223a-4157-9137-88dc492f2db2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e00f1c5-5c15-430f-baff-cf464bf5d9b3', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 826.439894] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Creating folder: Project (ea161b1af2bb4b9f8cc39b16582f7013). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 826.440352] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-055ce4a0-db87-4741-9848-c10df26b65cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.451256] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Created folder: Project (ea161b1af2bb4b9f8cc39b16582f7013) in parent group-v282025. [ 826.451634] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Creating folder: Instances. Parent ref: group-v282069. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 826.451985] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94a3d25d-b33a-4fc4-bf1a-d82269af4298 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.463014] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Created folder: Instances in parent group-v282069. [ 826.463014] env[62521]: DEBUG oslo.service.loopingcall [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.463014] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 826.463014] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc0fdfea-0786-4188-bd9a-c7b3438fd2d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.483811] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 826.483811] env[62521]: value = "task-1318493" [ 826.483811] env[62521]: _type = "Task" [ 826.483811] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.492158] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318493, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.526280] env[62521]: DEBUG nova.network.neutron [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Successfully created port: 12ed6066-6ad4-42ab-93ff-b25c2624d917 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.546024] env[62521]: DEBUG nova.compute.manager [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 826.549339] env[62521]: DEBUG nova.compute.manager [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.598148] env[62521]: DEBUG nova.compute.manager [req-c6746477-f616-4e6c-a25a-f6e5abd2a659 req-836f5632-0c2d-4628-9ae9-4dd3ff3b29fc service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Received event network-changed-7e00f1c5-5c15-430f-baff-cf464bf5d9b3 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 826.598560] env[62521]: DEBUG nova.compute.manager [req-c6746477-f616-4e6c-a25a-f6e5abd2a659 req-836f5632-0c2d-4628-9ae9-4dd3ff3b29fc service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Refreshing instance network info cache due to event network-changed-7e00f1c5-5c15-430f-baff-cf464bf5d9b3. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 826.600040] env[62521]: DEBUG oslo_concurrency.lockutils [req-c6746477-f616-4e6c-a25a-f6e5abd2a659 req-836f5632-0c2d-4628-9ae9-4dd3ff3b29fc service nova] Acquiring lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.600040] env[62521]: DEBUG oslo_concurrency.lockutils [req-c6746477-f616-4e6c-a25a-f6e5abd2a659 req-836f5632-0c2d-4628-9ae9-4dd3ff3b29fc service nova] Acquired lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.600040] env[62521]: DEBUG nova.network.neutron [req-c6746477-f616-4e6c-a25a-f6e5abd2a659 req-836f5632-0c2d-4628-9ae9-4dd3ff3b29fc service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Refreshing network info cache for port 7e00f1c5-5c15-430f-baff-cf464bf5d9b3 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 826.703277] env[62521]: DEBUG oslo_concurrency.lockutils [req-1c08bda8-fbb5-44d0-ad65-e0e9ff87b8e9 req-a1e1bb04-0694-4339-b3d2-12f0dffb56a7 service nova] Releasing lock "refresh_cache-af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.766220] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52445020-80d1-8fdb-3db4-9281f48766c5, 'name': SearchDatastore_Task, 'duration_secs': 0.01377} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.768071] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d92dabdd-90de-4f03-a356-d0d0b5f73a66 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.778889] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 826.778889] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52708fb0-f8a0-7f30-37db-f78894838112" [ 826.778889] env[62521]: _type = "Task" [ 826.778889] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.789962] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52708fb0-f8a0-7f30-37db-f78894838112, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.811163] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ebe858-81e3-4611-972b-b115e6fb301f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.819268] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73a2131-68f5-4c5f-ba14-b78df3770e78 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.854878] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d268059-fcbd-4446-b43b-346bd1e83dbe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.863123] env[62521]: DEBUG oslo_vmware.api [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318490, 'name': PowerOnVM_Task, 'duration_secs': 0.890581} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.865272] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 826.865487] env[62521]: INFO nova.compute.manager [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Took 9.42 seconds to spawn the instance on the hypervisor. [ 826.865691] env[62521]: DEBUG nova.compute.manager [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.869980] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8480621c-328b-4cc0-b0e1-2654873c68d2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.874414] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e96d802-49b9-47a1-8adc-e88cb177f520 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.897817] env[62521]: DEBUG nova.compute.provider_tree [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.994779] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318493, 'name': CreateVM_Task, 'duration_secs': 0.484599} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.994969] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 826.995708] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.995878] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.996195] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.996456] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed13f3ef-7dad-48b8-b5f8-6a7e0226f89d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.001351] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 827.001351] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a8e1c9-fc18-4b17-26ee-4cbb85ddbf0b" [ 827.001351] env[62521]: _type = "Task" [ 827.001351] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.009547] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a8e1c9-fc18-4b17-26ee-4cbb85ddbf0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.090726] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.292200] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52708fb0-f8a0-7f30-37db-f78894838112, 'name': SearchDatastore_Task, 'duration_secs': 0.010489} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.294617] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.294991] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] af91de53-2ef8-48d1-b8c1-7bd0809f4ae4/af91de53-2ef8-48d1-b8c1-7bd0809f4ae4.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.295188] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1a688d6-8983-4fb8-b2a7-f898a0aa6c40 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.302324] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 827.302324] env[62521]: value = "task-1318494" [ 827.302324] env[62521]: _type = "Task" [ 827.302324] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.310534] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318494, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.330638] env[62521]: DEBUG nova.network.neutron [req-c6746477-f616-4e6c-a25a-f6e5abd2a659 req-836f5632-0c2d-4628-9ae9-4dd3ff3b29fc service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updated VIF entry in instance network info cache for port 7e00f1c5-5c15-430f-baff-cf464bf5d9b3. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 827.331011] env[62521]: DEBUG nova.network.neutron [req-c6746477-f616-4e6c-a25a-f6e5abd2a659 req-836f5632-0c2d-4628-9ae9-4dd3ff3b29fc service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance_info_cache with network_info: [{"id": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "address": "fa:16:3e:d7:52:1f", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e00f1c5-5c", "ovs_interfaceid": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.404934] env[62521]: DEBUG nova.scheduler.client.report [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.412828] env[62521]: INFO nova.compute.manager [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Took 38.28 seconds to build instance. [ 827.511586] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a8e1c9-fc18-4b17-26ee-4cbb85ddbf0b, 'name': SearchDatastore_Task, 'duration_secs': 0.012018} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.511895] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.512147] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 827.512380] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.512552] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.512746] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 827.512945] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21cf7e44-d413-453e-b97c-6e31d150446c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.520564] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 827.520641] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 827.521290] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68ef5644-eff4-4b93-bcb5-d8038b13e011 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.526117] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 827.526117] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52335883-e774-f707-ecd9-7fe66b0f799c" [ 827.526117] env[62521]: _type = "Task" [ 827.526117] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.533421] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52335883-e774-f707-ecd9-7fe66b0f799c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.555896] env[62521]: DEBUG nova.compute.manager [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.581287] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.581544] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.581704] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.581890] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.582051] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.582204] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.582411] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.582569] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.582734] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.582895] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.583294] env[62521]: DEBUG nova.virt.hardware [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.584043] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ce8550-d08a-44a2-946d-78b449ce616f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.592670] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbbbc2a8-2aea-4a4a-a1ea-e0a75a59954c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.812227] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318494, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.834200] env[62521]: DEBUG oslo_concurrency.lockutils [req-c6746477-f616-4e6c-a25a-f6e5abd2a659 req-836f5632-0c2d-4628-9ae9-4dd3ff3b29fc service nova] Releasing lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.912029] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.912029] env[62521]: DEBUG nova.compute.manager [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.913231] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.072s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.913857] env[62521]: DEBUG nova.objects.instance [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lazy-loading 'resources' on Instance uuid 868c7798-8d0f-4c31-88dc-5007454796f1 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.916024] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b17424c9-e45a-46e6-b8ce-93b0438e9fa2 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "b3e33bab-636f-42df-ac6c-667464312c11" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.762s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.036988] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52335883-e774-f707-ecd9-7fe66b0f799c, 'name': SearchDatastore_Task, 'duration_secs': 0.011086} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.037946] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66b7b47c-3c88-4942-ab75-9cd1021d76e2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.043931] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 828.043931] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522979c6-973f-22cf-853e-9e320f71e877" [ 828.043931] env[62521]: _type = "Task" [ 828.043931] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.054234] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522979c6-973f-22cf-853e-9e320f71e877, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.313312] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318494, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.00516} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.313594] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] af91de53-2ef8-48d1-b8c1-7bd0809f4ae4/af91de53-2ef8-48d1-b8c1-7bd0809f4ae4.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.313820] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.314090] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-766a63d7-3e53-4964-ab78-6db3de24ec8a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.320893] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 828.320893] env[62521]: value = "task-1318495" [ 828.320893] env[62521]: _type = "Task" [ 828.320893] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.329612] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318495, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.335913] env[62521]: DEBUG nova.compute.manager [req-79180637-e92f-4b2b-8f76-5d20d00fc2a1 req-c2ac00ff-40c9-4558-bc81-8d7bcbedb905 service nova] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Received event network-vif-plugged-12ed6066-6ad4-42ab-93ff-b25c2624d917 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.336141] env[62521]: DEBUG oslo_concurrency.lockutils [req-79180637-e92f-4b2b-8f76-5d20d00fc2a1 req-c2ac00ff-40c9-4558-bc81-8d7bcbedb905 service nova] Acquiring lock "4dcd1c6c-8726-42db-997a-e78ba1293310-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.336500] env[62521]: DEBUG oslo_concurrency.lockutils [req-79180637-e92f-4b2b-8f76-5d20d00fc2a1 req-c2ac00ff-40c9-4558-bc81-8d7bcbedb905 service nova] Lock "4dcd1c6c-8726-42db-997a-e78ba1293310-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.336500] env[62521]: DEBUG oslo_concurrency.lockutils [req-79180637-e92f-4b2b-8f76-5d20d00fc2a1 req-c2ac00ff-40c9-4558-bc81-8d7bcbedb905 service nova] Lock "4dcd1c6c-8726-42db-997a-e78ba1293310-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.336693] env[62521]: DEBUG nova.compute.manager [req-79180637-e92f-4b2b-8f76-5d20d00fc2a1 req-c2ac00ff-40c9-4558-bc81-8d7bcbedb905 service nova] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] No waiting events found dispatching network-vif-plugged-12ed6066-6ad4-42ab-93ff-b25c2624d917 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.336872] env[62521]: WARNING nova.compute.manager [req-79180637-e92f-4b2b-8f76-5d20d00fc2a1 req-c2ac00ff-40c9-4558-bc81-8d7bcbedb905 service nova] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Received unexpected event network-vif-plugged-12ed6066-6ad4-42ab-93ff-b25c2624d917 for instance with vm_state building and task_state spawning. [ 828.416641] env[62521]: DEBUG nova.compute.utils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.420719] env[62521]: DEBUG nova.compute.manager [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 828.424014] env[62521]: DEBUG nova.network.neutron [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 828.424014] env[62521]: DEBUG nova.compute.manager [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.484068] env[62521]: DEBUG nova.policy [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d7aab6f3f6245c2b56979677428a9df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e6baab1556e4b00b07ab554664b4445', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 828.520192] env[62521]: DEBUG nova.network.neutron [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Successfully updated port: 12ed6066-6ad4-42ab-93ff-b25c2624d917 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.555805] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522979c6-973f-22cf-853e-9e320f71e877, 'name': SearchDatastore_Task, 'duration_secs': 0.058779} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.558476] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.558476] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] d5f6492d-ca1b-4e74-b792-b04d55c33660/d5f6492d-ca1b-4e74-b792-b04d55c33660.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 828.558953] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d2ca79e-8bf0-4704-ab91-d7bb957de27c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.566894] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 828.566894] env[62521]: value = "task-1318496" [ 828.566894] env[62521]: _type = "Task" [ 828.566894] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.577287] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318496, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.689022] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5cd978-58b0-4ed2-93b3-c04c046e47b3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.695856] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6034cf-2efe-46b1-81d8-17d591a16bfd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.729027] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84080d57-5399-4a97-870f-23a2d02045fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.736447] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b567a8-d269-405a-ae44-c81991d50e4f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.751973] env[62521]: DEBUG nova.compute.provider_tree [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.762930] env[62521]: DEBUG nova.compute.manager [req-bec2966b-9a11-4b38-9a9a-89e62a6ad364 req-f249e3e8-a6da-4206-a33b-c7a10227b07f service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Received event network-changed-53471722-7f78-46af-bc13-02c61f597490 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.763144] env[62521]: DEBUG nova.compute.manager [req-bec2966b-9a11-4b38-9a9a-89e62a6ad364 req-f249e3e8-a6da-4206-a33b-c7a10227b07f service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Refreshing instance network info cache due to event network-changed-53471722-7f78-46af-bc13-02c61f597490. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 828.763446] env[62521]: DEBUG oslo_concurrency.lockutils [req-bec2966b-9a11-4b38-9a9a-89e62a6ad364 req-f249e3e8-a6da-4206-a33b-c7a10227b07f service nova] Acquiring lock "refresh_cache-e8cabcdc-d728-434d-a40d-fe751db086c3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.763612] env[62521]: DEBUG oslo_concurrency.lockutils [req-bec2966b-9a11-4b38-9a9a-89e62a6ad364 req-f249e3e8-a6da-4206-a33b-c7a10227b07f service nova] Acquired lock "refresh_cache-e8cabcdc-d728-434d-a40d-fe751db086c3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.763783] env[62521]: DEBUG nova.network.neutron [req-bec2966b-9a11-4b38-9a9a-89e62a6ad364 req-f249e3e8-a6da-4206-a33b-c7a10227b07f service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Refreshing network info cache for port 53471722-7f78-46af-bc13-02c61f597490 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.831285] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318495, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.240414} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.831920] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.832716] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-631bc1c0-7cc4-4fac-8199-35b9222063d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.855426] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] af91de53-2ef8-48d1-b8c1-7bd0809f4ae4/af91de53-2ef8-48d1-b8c1-7bd0809f4ae4.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.856055] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71f3dc2e-2d0a-4d27-a5b5-a77f5578119a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.875241] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 828.875241] env[62521]: value = "task-1318497" [ 828.875241] env[62521]: _type = "Task" [ 828.875241] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.883628] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318497, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.931662] env[62521]: DEBUG nova.compute.manager [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.940660] env[62521]: DEBUG nova.network.neutron [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Successfully created port: a7d4852b-d44e-448f-b25d-d4477b2c3ed9 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.952853] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.023170] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "refresh_cache-4dcd1c6c-8726-42db-997a-e78ba1293310" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.023497] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "refresh_cache-4dcd1c6c-8726-42db-997a-e78ba1293310" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.023736] env[62521]: DEBUG nova.network.neutron [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 829.076745] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318496, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.255205] env[62521]: DEBUG nova.scheduler.client.report [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.400680] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318497, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.579272] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318496, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.614145] env[62521]: DEBUG nova.network.neutron [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.760811] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.847s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.764247] env[62521]: DEBUG nova.network.neutron [req-bec2966b-9a11-4b38-9a9a-89e62a6ad364 req-f249e3e8-a6da-4206-a33b-c7a10227b07f service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Updated VIF entry in instance network info cache for port 53471722-7f78-46af-bc13-02c61f597490. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 829.764580] env[62521]: DEBUG nova.network.neutron [req-bec2966b-9a11-4b38-9a9a-89e62a6ad364 req-f249e3e8-a6da-4206-a33b-c7a10227b07f service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Updating instance_info_cache with network_info: [{"id": "53471722-7f78-46af-bc13-02c61f597490", "address": "fa:16:3e:8f:1a:f8", "network": {"id": "3d22562d-07d9-4b37-be52-90908a15d111", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-61847516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a207fb2d461340c4b2fc6e456596d719", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53471722-7f", "ovs_interfaceid": "53471722-7f78-46af-bc13-02c61f597490", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.765795] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.574s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.767324] env[62521]: INFO nova.compute.claims [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.783060] env[62521]: INFO nova.scheduler.client.report [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Deleted allocations for instance 868c7798-8d0f-4c31-88dc-5007454796f1 [ 829.887226] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318497, 'name': ReconfigVM_Task, 'duration_secs': 0.756531} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.887526] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Reconfigured VM instance instance-0000003a to attach disk [datastore1] af91de53-2ef8-48d1-b8c1-7bd0809f4ae4/af91de53-2ef8-48d1-b8c1-7bd0809f4ae4.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.888225] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-676040a5-900c-41a8-bdd1-b03962574048 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.895759] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 829.895759] env[62521]: value = "task-1318498" [ 829.895759] env[62521]: _type = "Task" [ 829.895759] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.905076] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318498, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.948989] env[62521]: DEBUG nova.compute.manager [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.981391] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.981766] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.981995] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.982273] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.982492] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.982708] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.982952] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.983147] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.983325] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.983557] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.983736] env[62521]: DEBUG nova.virt.hardware [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.984963] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c796dcdd-7af6-43f4-9c4f-2c9af253a1ef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.993732] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4582a5e8-4832-45bb-acba-4c926a308f60 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.078204] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318496, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.097023} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.078535] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] d5f6492d-ca1b-4e74-b792-b04d55c33660/d5f6492d-ca1b-4e74-b792-b04d55c33660.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 830.078764] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.079131] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0703e26e-ca3c-41c6-9e81-6d639ff70105 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.085962] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 830.085962] env[62521]: value = "task-1318499" [ 830.085962] env[62521]: _type = "Task" [ 830.085962] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.093868] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318499, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.100111] env[62521]: DEBUG nova.network.neutron [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Updating instance_info_cache with network_info: [{"id": "12ed6066-6ad4-42ab-93ff-b25c2624d917", "address": "fa:16:3e:44:57:d2", "network": {"id": "ce7ccdc7-e99d-48bd-93a9-c8c14725d58b", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1698306226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25420ed4f82e478989a011d40f037a24", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b56036cd-97ac-47f5-9089-7b38bfe99228", "external-id": "nsx-vlan-transportzone-301", "segmentation_id": 301, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ed6066-6a", "ovs_interfaceid": "12ed6066-6ad4-42ab-93ff-b25c2624d917", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.271243] env[62521]: DEBUG oslo_concurrency.lockutils [req-bec2966b-9a11-4b38-9a9a-89e62a6ad364 req-f249e3e8-a6da-4206-a33b-c7a10227b07f service nova] Releasing lock "refresh_cache-e8cabcdc-d728-434d-a40d-fe751db086c3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.292935] env[62521]: DEBUG oslo_concurrency.lockutils [None req-540607e8-7fea-4eaa-a0b2-5e9e4c141dc4 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "868c7798-8d0f-4c31-88dc-5007454796f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.212s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.361889] env[62521]: DEBUG nova.compute.manager [req-43ba852e-f706-4540-abdf-c1669c1d2175 req-e13ce39f-96b3-43dd-996e-10b6e4adc2d2 service nova] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Received event network-changed-12ed6066-6ad4-42ab-93ff-b25c2624d917 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.362091] env[62521]: DEBUG nova.compute.manager [req-43ba852e-f706-4540-abdf-c1669c1d2175 req-e13ce39f-96b3-43dd-996e-10b6e4adc2d2 service nova] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Refreshing instance network info cache due to event network-changed-12ed6066-6ad4-42ab-93ff-b25c2624d917. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 830.362381] env[62521]: DEBUG oslo_concurrency.lockutils [req-43ba852e-f706-4540-abdf-c1669c1d2175 req-e13ce39f-96b3-43dd-996e-10b6e4adc2d2 service nova] Acquiring lock "refresh_cache-4dcd1c6c-8726-42db-997a-e78ba1293310" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.371402] env[62521]: DEBUG nova.compute.manager [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.372265] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09719592-ea9d-4aff-84f4-83b76efd8650 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.404806] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318498, 'name': Rename_Task, 'duration_secs': 0.163327} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.404946] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 830.405142] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e0c9491-b9ca-44f4-8365-a0e227332b61 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.412158] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 830.412158] env[62521]: value = "task-1318500" [ 830.412158] env[62521]: _type = "Task" [ 830.412158] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.420076] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318500, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.451267] env[62521]: DEBUG nova.network.neutron [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Successfully updated port: a7d4852b-d44e-448f-b25d-d4477b2c3ed9 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 830.596164] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318499, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071145} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.596511] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 830.597379] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ce2cf3-1c16-43b1-a57d-7c0a7b2c5d33 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.602732] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "refresh_cache-4dcd1c6c-8726-42db-997a-e78ba1293310" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.603036] env[62521]: DEBUG nova.compute.manager [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Instance network_info: |[{"id": "12ed6066-6ad4-42ab-93ff-b25c2624d917", "address": "fa:16:3e:44:57:d2", "network": {"id": "ce7ccdc7-e99d-48bd-93a9-c8c14725d58b", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1698306226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25420ed4f82e478989a011d40f037a24", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b56036cd-97ac-47f5-9089-7b38bfe99228", "external-id": "nsx-vlan-transportzone-301", "segmentation_id": 301, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ed6066-6a", "ovs_interfaceid": "12ed6066-6ad4-42ab-93ff-b25c2624d917", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.615230] env[62521]: DEBUG oslo_concurrency.lockutils [req-43ba852e-f706-4540-abdf-c1669c1d2175 req-e13ce39f-96b3-43dd-996e-10b6e4adc2d2 service nova] Acquired lock "refresh_cache-4dcd1c6c-8726-42db-997a-e78ba1293310" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.615445] env[62521]: DEBUG nova.network.neutron [req-43ba852e-f706-4540-abdf-c1669c1d2175 req-e13ce39f-96b3-43dd-996e-10b6e4adc2d2 service nova] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Refreshing network info cache for port 12ed6066-6ad4-42ab-93ff-b25c2624d917 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.617253] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:57:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b56036cd-97ac-47f5-9089-7b38bfe99228', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12ed6066-6ad4-42ab-93ff-b25c2624d917', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.624886] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Creating folder: Project (25420ed4f82e478989a011d40f037a24). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.633799] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] d5f6492d-ca1b-4e74-b792-b04d55c33660/d5f6492d-ca1b-4e74-b792-b04d55c33660.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.635347] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da0e25d7-e2a7-4dbb-af73-d8488b7f9be8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.637370] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48a42750-9974-4ff2-9329-4bc0d7cb1017 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.657043] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 830.657043] env[62521]: value = "task-1318502" [ 830.657043] env[62521]: _type = "Task" [ 830.657043] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.663852] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Created folder: Project (25420ed4f82e478989a011d40f037a24) in parent group-v282025. [ 830.664053] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Creating folder: Instances. Parent ref: group-v282072. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.667331] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2fc59e7c-8af4-46b5-bda8-9056f8e25eb2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.668815] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318502, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.679095] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Created folder: Instances in parent group-v282072. [ 830.679341] env[62521]: DEBUG oslo.service.loopingcall [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.679561] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.679772] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-968eed54-5174-4186-9470-7506b2dda921 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.700291] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.700291] env[62521]: value = "task-1318504" [ 830.700291] env[62521]: _type = "Task" [ 830.700291] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.707426] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318504, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.789264] env[62521]: DEBUG nova.compute.manager [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Received event network-vif-plugged-a7d4852b-d44e-448f-b25d-d4477b2c3ed9 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.789608] env[62521]: DEBUG oslo_concurrency.lockutils [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] Acquiring lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.789937] env[62521]: DEBUG oslo_concurrency.lockutils [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] Lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.790233] env[62521]: DEBUG oslo_concurrency.lockutils [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] Lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.790523] env[62521]: DEBUG nova.compute.manager [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] No waiting events found dispatching network-vif-plugged-a7d4852b-d44e-448f-b25d-d4477b2c3ed9 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 830.790818] env[62521]: WARNING nova.compute.manager [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Received unexpected event network-vif-plugged-a7d4852b-d44e-448f-b25d-d4477b2c3ed9 for instance with vm_state building and task_state spawning. [ 830.791131] env[62521]: DEBUG nova.compute.manager [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Received event network-changed-a7d4852b-d44e-448f-b25d-d4477b2c3ed9 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.791415] env[62521]: DEBUG nova.compute.manager [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Refreshing instance network info cache due to event network-changed-a7d4852b-d44e-448f-b25d-d4477b2c3ed9. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 830.791724] env[62521]: DEBUG oslo_concurrency.lockutils [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] Acquiring lock "refresh_cache-70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.791973] env[62521]: DEBUG oslo_concurrency.lockutils [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] Acquired lock "refresh_cache-70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.792266] env[62521]: DEBUG nova.network.neutron [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Refreshing network info cache for port a7d4852b-d44e-448f-b25d-d4477b2c3ed9 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.883032] env[62521]: INFO nova.compute.manager [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] instance snapshotting [ 830.886069] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce6ec31-36d7-4078-9e82-c97d369fa0a5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.925308] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f58eae7-195b-4557-8a42-21bf59cf6b9e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.938717] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318500, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.956651] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.031733] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8afef5c1-bd7c-484a-ae09-22206081c69b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.040031] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fbb9f4-9af9-488b-9f01-fb190c707049 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.076490] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce5f7ee-215a-4fd8-8710-04df40bc1a59 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.084192] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fd7096-be6e-42cc-b3d1-5670222bb7c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.098872] env[62521]: DEBUG nova.compute.provider_tree [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.107333] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "062f311c-a783-41cf-8d83-b3a8d4df14f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.107565] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "062f311c-a783-41cf-8d83-b3a8d4df14f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.107757] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "062f311c-a783-41cf-8d83-b3a8d4df14f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.107943] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "062f311c-a783-41cf-8d83-b3a8d4df14f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.108302] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "062f311c-a783-41cf-8d83-b3a8d4df14f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.110975] env[62521]: INFO nova.compute.manager [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Terminating instance [ 831.112768] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "refresh_cache-062f311c-a783-41cf-8d83-b3a8d4df14f4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.112979] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquired lock "refresh_cache-062f311c-a783-41cf-8d83-b3a8d4df14f4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.113190] env[62521]: DEBUG nova.network.neutron [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.167214] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318502, 'name': ReconfigVM_Task, 'duration_secs': 0.385401} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.167319] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Reconfigured VM instance instance-0000003b to attach disk [datastore1] d5f6492d-ca1b-4e74-b792-b04d55c33660/d5f6492d-ca1b-4e74-b792-b04d55c33660.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.167941] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-06e3f2d2-236c-44e9-9c3a-bccd715b423c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.173961] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 831.173961] env[62521]: value = "task-1318505" [ 831.173961] env[62521]: _type = "Task" [ 831.173961] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.182190] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318505, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.210045] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318504, 'name': CreateVM_Task, 'duration_secs': 0.459848} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.211976] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 831.212666] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.212832] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.213164] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.213744] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15867c46-fddd-4144-b509-ba08bbc11414 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.218651] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 831.218651] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524d9eca-8bc8-71e0-fc5d-4f5a990e8abc" [ 831.218651] env[62521]: _type = "Task" [ 831.218651] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.226268] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524d9eca-8bc8-71e0-fc5d-4f5a990e8abc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.349391] env[62521]: DEBUG nova.network.neutron [req-43ba852e-f706-4540-abdf-c1669c1d2175 req-e13ce39f-96b3-43dd-996e-10b6e4adc2d2 service nova] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Updated VIF entry in instance network info cache for port 12ed6066-6ad4-42ab-93ff-b25c2624d917. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 831.349626] env[62521]: DEBUG nova.network.neutron [req-43ba852e-f706-4540-abdf-c1669c1d2175 req-e13ce39f-96b3-43dd-996e-10b6e4adc2d2 service nova] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Updating instance_info_cache with network_info: [{"id": "12ed6066-6ad4-42ab-93ff-b25c2624d917", "address": "fa:16:3e:44:57:d2", "network": {"id": "ce7ccdc7-e99d-48bd-93a9-c8c14725d58b", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1698306226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25420ed4f82e478989a011d40f037a24", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b56036cd-97ac-47f5-9089-7b38bfe99228", "external-id": "nsx-vlan-transportzone-301", "segmentation_id": 301, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ed6066-6a", "ovs_interfaceid": "12ed6066-6ad4-42ab-93ff-b25c2624d917", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.351484] env[62521]: DEBUG nova.network.neutron [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.427547] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318500, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.450868] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 831.451169] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1e2f48a6-77bc-475b-8730-017563fe7aac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.457899] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 831.457899] env[62521]: value = "task-1318506" [ 831.457899] env[62521]: _type = "Task" [ 831.457899] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.458689] env[62521]: DEBUG nova.network.neutron [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.468745] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318506, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.602380] env[62521]: DEBUG nova.scheduler.client.report [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.642658] env[62521]: DEBUG nova.network.neutron [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.683590] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318505, 'name': Rename_Task, 'duration_secs': 0.186632} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.683866] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 831.684138] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-075977b9-f836-48a2-8b28-46825b633c2f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.690595] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 831.690595] env[62521]: value = "task-1318507" [ 831.690595] env[62521]: _type = "Task" [ 831.690595] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.698801] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318507, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.731042] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524d9eca-8bc8-71e0-fc5d-4f5a990e8abc, 'name': SearchDatastore_Task, 'duration_secs': 0.009775} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.731042] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.731042] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.731042] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.731174] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.731174] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.731174] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64ada54f-404d-4d5b-bf5d-70278129700d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.733060] env[62521]: DEBUG nova.network.neutron [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.740528] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.740737] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 831.741486] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b0d0845-9531-449c-adb4-fdbd30d7f37f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.747373] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 831.747373] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529f33ae-1400-9c2f-22d5-8c9a83d4ad54" [ 831.747373] env[62521]: _type = "Task" [ 831.747373] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.756407] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529f33ae-1400-9c2f-22d5-8c9a83d4ad54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.857220] env[62521]: DEBUG oslo_concurrency.lockutils [req-43ba852e-f706-4540-abdf-c1669c1d2175 req-e13ce39f-96b3-43dd-996e-10b6e4adc2d2 service nova] Releasing lock "refresh_cache-4dcd1c6c-8726-42db-997a-e78ba1293310" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.927398] env[62521]: DEBUG oslo_vmware.api [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318500, 'name': PowerOnVM_Task, 'duration_secs': 1.32702} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.927708] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 831.927975] env[62521]: INFO nova.compute.manager [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Took 10.16 seconds to spawn the instance on the hypervisor. [ 831.928251] env[62521]: DEBUG nova.compute.manager [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.929116] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4f8413-90e7-4d03-9f87-bfd8c8bd4662 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.966504] env[62521]: DEBUG oslo_concurrency.lockutils [req-9973e89f-1267-4893-a544-2370ad0efeb8 req-69c483e3-e721-468d-b56f-c0b7cf5cd388 service nova] Releasing lock "refresh_cache-70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.967104] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.967372] env[62521]: DEBUG nova.network.neutron [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.977483] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318506, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.108297] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.108870] env[62521]: DEBUG nova.compute.manager [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.111776] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.082s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.113595] env[62521]: INFO nova.compute.claims [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.200990] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318507, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.235547] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Releasing lock "refresh_cache-062f311c-a783-41cf-8d83-b3a8d4df14f4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.236120] env[62521]: DEBUG nova.compute.manager [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.236342] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.237494] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-734902c3-81b9-4a31-a121-e9c115389688 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.245499] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.245796] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c931ef75-4e00-4239-ab19-162964e6820a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.252758] env[62521]: DEBUG oslo_vmware.api [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 832.252758] env[62521]: value = "task-1318508" [ 832.252758] env[62521]: _type = "Task" [ 832.252758] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.256774] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529f33ae-1400-9c2f-22d5-8c9a83d4ad54, 'name': SearchDatastore_Task, 'duration_secs': 0.01073} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.260499] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-064aca64-6220-4cbc-b54f-edd7dacc2287 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.265825] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 832.265825] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523bee4d-31a6-c3b1-392e-7c7707db10c9" [ 832.265825] env[62521]: _type = "Task" [ 832.265825] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.269082] env[62521]: DEBUG oslo_vmware.api [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318508, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.276647] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523bee4d-31a6-c3b1-392e-7c7707db10c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.447252] env[62521]: INFO nova.compute.manager [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Took 38.94 seconds to build instance. [ 832.472625] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318506, 'name': CreateSnapshot_Task, 'duration_secs': 0.676161} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.472894] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 832.473948] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488add5d-6c93-47fe-a4e6-61e92baca4ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.521238] env[62521]: DEBUG nova.network.neutron [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.618117] env[62521]: DEBUG nova.compute.utils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.624109] env[62521]: DEBUG nova.compute.manager [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.624109] env[62521]: DEBUG nova.network.neutron [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 832.697555] env[62521]: DEBUG nova.policy [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0623f9310e84db7abb41e0ae57c1b76', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f87a25d6e31a47e397ebb9c6aa527305', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.704986] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318507, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.740506] env[62521]: DEBUG nova.network.neutron [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Updating instance_info_cache with network_info: [{"id": "a7d4852b-d44e-448f-b25d-d4477b2c3ed9", "address": "fa:16:3e:05:86:09", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7d4852b-d4", "ovs_interfaceid": "a7d4852b-d44e-448f-b25d-d4477b2c3ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.767934] env[62521]: DEBUG oslo_vmware.api [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318508, 'name': PowerOffVM_Task, 'duration_secs': 0.173132} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.768358] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.768536] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.768793] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cfef9e8c-1b3a-41dd-bcaa-7b693650773b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.780991] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523bee4d-31a6-c3b1-392e-7c7707db10c9, 'name': SearchDatastore_Task, 'duration_secs': 0.020372} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.781255] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.781505] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 4dcd1c6c-8726-42db-997a-e78ba1293310/4dcd1c6c-8726-42db-997a-e78ba1293310.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 832.781753] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0dc5783a-2630-47a9-b8d6-a2ee4d8be0f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.787797] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 832.787797] env[62521]: value = "task-1318510" [ 832.787797] env[62521]: _type = "Task" [ 832.787797] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.797438] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318510, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.798702] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 832.798899] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 832.799233] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Deleting the datastore file [datastore1] 062f311c-a783-41cf-8d83-b3a8d4df14f4 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.799356] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a167fe3-2929-4234-b235-5012570e3d11 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.805874] env[62521]: DEBUG oslo_vmware.api [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for the task: (returnval){ [ 832.805874] env[62521]: value = "task-1318511" [ 832.805874] env[62521]: _type = "Task" [ 832.805874] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.813479] env[62521]: DEBUG oslo_vmware.api [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318511, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.948954] env[62521]: DEBUG oslo_concurrency.lockutils [None req-10c84731-525c-49e1-be98-b046d3e9f877 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.539s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.992142] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 832.992493] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-08b03169-1ced-4999-acba-4ec0311a2583 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.001739] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 833.001739] env[62521]: value = "task-1318512" [ 833.001739] env[62521]: _type = "Task" [ 833.001739] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.009937] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318512, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.122878] env[62521]: DEBUG nova.compute.manager [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.191028] env[62521]: DEBUG nova.network.neutron [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Successfully created port: b520981a-8f32-4f58-9e84-a062b20d933a {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.203903] env[62521]: DEBUG oslo_vmware.api [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318507, 'name': PowerOnVM_Task, 'duration_secs': 1.101367} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.206820] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.209717] env[62521]: INFO nova.compute.manager [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Took 9.07 seconds to spawn the instance on the hypervisor. [ 833.209973] env[62521]: DEBUG nova.compute.manager [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.211068] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c6ba76-eec5-4112-b6df-65c92b8e1c6c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.243092] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.243432] env[62521]: DEBUG nova.compute.manager [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Instance network_info: |[{"id": "a7d4852b-d44e-448f-b25d-d4477b2c3ed9", "address": "fa:16:3e:05:86:09", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7d4852b-d4", "ovs_interfaceid": "a7d4852b-d44e-448f-b25d-d4477b2c3ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.244473] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:86:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7d4852b-d44e-448f-b25d-d4477b2c3ed9', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.254205] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Creating folder: Project (7e6baab1556e4b00b07ab554664b4445). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.257773] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fcb9b8d5-6aa0-4cfd-8d30-d6b3030b1aaa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.269077] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Created folder: Project (7e6baab1556e4b00b07ab554664b4445) in parent group-v282025. [ 833.269316] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Creating folder: Instances. Parent ref: group-v282077. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.269573] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-edf83144-79de-4f12-be0e-edf214561bc0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.280201] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Created folder: Instances in parent group-v282077. [ 833.281046] env[62521]: DEBUG oslo.service.loopingcall [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.281046] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.281046] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-998ca68c-be9f-4018-be95-92808a84391b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.305017] env[62521]: DEBUG nova.compute.manager [req-819498b9-d7c9-4bd5-987a-8e6470decdd7 req-2ce595ba-8488-4bef-a915-a9256d3da336 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Received event network-changed-6cf47748-b470-4bc0-8a01-23c0154e45be {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.305237] env[62521]: DEBUG nova.compute.manager [req-819498b9-d7c9-4bd5-987a-8e6470decdd7 req-2ce595ba-8488-4bef-a915-a9256d3da336 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Refreshing instance network info cache due to event network-changed-6cf47748-b470-4bc0-8a01-23c0154e45be. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.305457] env[62521]: DEBUG oslo_concurrency.lockutils [req-819498b9-d7c9-4bd5-987a-8e6470decdd7 req-2ce595ba-8488-4bef-a915-a9256d3da336 service nova] Acquiring lock "refresh_cache-af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.305661] env[62521]: DEBUG oslo_concurrency.lockutils [req-819498b9-d7c9-4bd5-987a-8e6470decdd7 req-2ce595ba-8488-4bef-a915-a9256d3da336 service nova] Acquired lock "refresh_cache-af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.305852] env[62521]: DEBUG nova.network.neutron [req-819498b9-d7c9-4bd5-987a-8e6470decdd7 req-2ce595ba-8488-4bef-a915-a9256d3da336 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Refreshing network info cache for port 6cf47748-b470-4bc0-8a01-23c0154e45be {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.312815] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.312815] env[62521]: value = "task-1318515" [ 833.312815] env[62521]: _type = "Task" [ 833.312815] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.321299] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318510, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.328395] env[62521]: DEBUG oslo_vmware.api [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Task: {'id': task-1318511, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.271327} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.329271] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.329402] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 833.329573] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 833.329769] env[62521]: INFO nova.compute.manager [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Took 1.09 seconds to destroy the instance on the hypervisor. [ 833.330098] env[62521]: DEBUG oslo.service.loopingcall [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.334206] env[62521]: DEBUG nova.compute.manager [-] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 833.334298] env[62521]: DEBUG nova.network.neutron [-] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 833.336183] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318515, 'name': CreateVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.371874] env[62521]: DEBUG nova.network.neutron [-] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.452862] env[62521]: DEBUG nova.compute.manager [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 833.490446] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fdb5407-3378-4964-89b0-e5053e98dc60 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.500586] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c022bdb-72ce-44ab-aa2b-02dc727e4b43 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.543927] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318512, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.546930] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a009a9b-a497-426f-a972-5e4a4f975b4c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.558672] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a03a959-e8a3-4bb9-b622-e0b0468a005d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.574126] env[62521]: DEBUG nova.compute.provider_tree [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.731751] env[62521]: INFO nova.compute.manager [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Took 38.22 seconds to build instance. [ 833.809023] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318510, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.843149} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.809023] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 4dcd1c6c-8726-42db-997a-e78ba1293310/4dcd1c6c-8726-42db-997a-e78ba1293310.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 833.809023] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 833.809241] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd5fb016-6309-4c95-bf51-04b397e2c247 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.818534] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 833.818534] env[62521]: value = "task-1318516" [ 833.818534] env[62521]: _type = "Task" [ 833.818534] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.833974] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318516, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.837128] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318515, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.877020] env[62521]: DEBUG nova.network.neutron [-] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.977787] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.015216] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318512, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.078294] env[62521]: DEBUG nova.scheduler.client.report [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.119829] env[62521]: DEBUG nova.network.neutron [req-819498b9-d7c9-4bd5-987a-8e6470decdd7 req-2ce595ba-8488-4bef-a915-a9256d3da336 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Updated VIF entry in instance network info cache for port 6cf47748-b470-4bc0-8a01-23c0154e45be. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 834.120192] env[62521]: DEBUG nova.network.neutron [req-819498b9-d7c9-4bd5-987a-8e6470decdd7 req-2ce595ba-8488-4bef-a915-a9256d3da336 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Updating instance_info_cache with network_info: [{"id": "6cf47748-b470-4bc0-8a01-23c0154e45be", "address": "fa:16:3e:4d:ae:61", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.134", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cf47748-b4", "ovs_interfaceid": "6cf47748-b470-4bc0-8a01-23c0154e45be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.134770] env[62521]: DEBUG nova.compute.manager [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.190113] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.190370] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.190526] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.190706] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.191342] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.191342] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.191342] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.191342] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.191537] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.191667] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.191836] env[62521]: DEBUG nova.virt.hardware [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.192704] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9ab0f6-d6c1-4aad-806e-88bcb0e512fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.200535] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d51851-baca-4184-a8bf-7aaef85d57db {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.233900] env[62521]: DEBUG oslo_concurrency.lockutils [None req-02eb17cc-c36a-4f7c-b460-3bb1041d551b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.161s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.331209] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318516, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069226} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.334507] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.335238] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318515, 'name': CreateVM_Task, 'duration_secs': 0.709263} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.335499] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714e0f24-9664-44a3-ad45-0e76aa89e559 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.337759] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.338538] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.338657] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.339069] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.339644] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6414f30-5998-4187-964b-1d2ecf943e2b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.358975] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 4dcd1c6c-8726-42db-997a-e78ba1293310/4dcd1c6c-8726-42db-997a-e78ba1293310.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.360372] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6cbe8171-54e1-45bd-9564-4909b36c68ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.375352] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 834.375352] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52584470-5eae-12d4-ab4e-254b79666241" [ 834.375352] env[62521]: _type = "Task" [ 834.375352] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.379530] env[62521]: INFO nova.compute.manager [-] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Took 1.05 seconds to deallocate network for instance. [ 834.385247] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 834.385247] env[62521]: value = "task-1318517" [ 834.385247] env[62521]: _type = "Task" [ 834.385247] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.388934] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52584470-5eae-12d4-ab4e-254b79666241, 'name': SearchDatastore_Task, 'duration_secs': 0.015755} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.392402] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.392577] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.392812] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.392961] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.393164] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.393419] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-acd3d7ff-4ccd-434a-8167-53dbf5408824 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.401161] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318517, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.402767] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.402767] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.403495] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11b020c0-b114-4f55-a0bc-a73074ccac5b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.408745] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 834.408745] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525352b6-6c1f-1db0-7f21-0f63685bd427" [ 834.408745] env[62521]: _type = "Task" [ 834.408745] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.416641] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525352b6-6c1f-1db0-7f21-0f63685bd427, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.515301] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318512, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.583240] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.583846] env[62521]: DEBUG nova.compute.manager [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.586587] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.792s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.586829] env[62521]: DEBUG nova.objects.instance [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lazy-loading 'resources' on Instance uuid 50aef23e-9429-4499-9fc1-c31b97e23029 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.622876] env[62521]: DEBUG oslo_concurrency.lockutils [req-819498b9-d7c9-4bd5-987a-8e6470decdd7 req-2ce595ba-8488-4bef-a915-a9256d3da336 service nova] Releasing lock "refresh_cache-af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.846259] env[62521]: DEBUG nova.network.neutron [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Successfully updated port: b520981a-8f32-4f58-9e84-a062b20d933a {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.894344] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.900824] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318517, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.919050] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525352b6-6c1f-1db0-7f21-0f63685bd427, 'name': SearchDatastore_Task, 'duration_secs': 0.011119} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.919875] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-434eacd8-5d64-47db-87a7-cc260a37c179 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.925643] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 834.925643] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52573d2e-278c-385e-4b3d-d53574e77fd4" [ 834.925643] env[62521]: _type = "Task" [ 834.925643] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.934109] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52573d2e-278c-385e-4b3d-d53574e77fd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.017161] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318512, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.090270] env[62521]: DEBUG nova.compute.utils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.091800] env[62521]: DEBUG nova.compute.manager [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 835.091927] env[62521]: DEBUG nova.network.neutron [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 835.152351] env[62521]: DEBUG nova.policy [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb1d6b613ddc4c8fa4c3bede87ffb00c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4facd01ddac841298d07810781c61e05', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 835.332971] env[62521]: DEBUG nova.compute.manager [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Received event network-vif-plugged-b520981a-8f32-4f58-9e84-a062b20d933a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.333396] env[62521]: DEBUG oslo_concurrency.lockutils [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] Acquiring lock "4afe09db-6c01-444f-a127-6e1f97794544-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.334312] env[62521]: DEBUG oslo_concurrency.lockutils [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] Lock "4afe09db-6c01-444f-a127-6e1f97794544-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.334312] env[62521]: DEBUG oslo_concurrency.lockutils [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] Lock "4afe09db-6c01-444f-a127-6e1f97794544-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.334312] env[62521]: DEBUG nova.compute.manager [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] No waiting events found dispatching network-vif-plugged-b520981a-8f32-4f58-9e84-a062b20d933a {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.334737] env[62521]: WARNING nova.compute.manager [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Received unexpected event network-vif-plugged-b520981a-8f32-4f58-9e84-a062b20d933a for instance with vm_state building and task_state spawning. [ 835.334831] env[62521]: DEBUG nova.compute.manager [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Received event network-changed-b520981a-8f32-4f58-9e84-a062b20d933a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.335185] env[62521]: DEBUG nova.compute.manager [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Refreshing instance network info cache due to event network-changed-b520981a-8f32-4f58-9e84-a062b20d933a. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 835.335349] env[62521]: DEBUG oslo_concurrency.lockutils [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] Acquiring lock "refresh_cache-4afe09db-6c01-444f-a127-6e1f97794544" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.335448] env[62521]: DEBUG oslo_concurrency.lockutils [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] Acquired lock "refresh_cache-4afe09db-6c01-444f-a127-6e1f97794544" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.335572] env[62521]: DEBUG nova.network.neutron [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Refreshing network info cache for port b520981a-8f32-4f58-9e84-a062b20d933a {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.339485] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9648f2ba-dfce-4100-a39f-db01d886a521 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.349074] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquiring lock "refresh_cache-4afe09db-6c01-444f-a127-6e1f97794544" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.352327] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c17b12c-695b-4d21-9fbd-5abf998cd651 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.403247] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95af504b-3e7b-430e-98ab-337bc19dbe07 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.414552] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318517, 'name': ReconfigVM_Task, 'duration_secs': 0.741213} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.415490] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 4dcd1c6c-8726-42db-997a-e78ba1293310/4dcd1c6c-8726-42db-997a-e78ba1293310.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.417402] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271fdc1a-9990-4984-8e16-09fc794f2d3e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.422025] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-485dbd1e-67a1-4c63-bed3-22d88d423225 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.440479] env[62521]: DEBUG nova.compute.provider_tree [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.442771] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 835.442771] env[62521]: value = "task-1318518" [ 835.442771] env[62521]: _type = "Task" [ 835.442771] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.443782] env[62521]: DEBUG nova.network.neutron [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Successfully created port: ef87a463-92f6-449b-8262-f156eaca1596 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.460155] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318518, 'name': Rename_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.463639] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52573d2e-278c-385e-4b3d-d53574e77fd4, 'name': SearchDatastore_Task, 'duration_secs': 0.010414} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.464242] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.464487] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6/70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.464790] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a94a94df-b120-4b5b-98fe-c29e1f27d869 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.471502] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 835.471502] env[62521]: value = "task-1318519" [ 835.471502] env[62521]: _type = "Task" [ 835.471502] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.480177] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318519, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.496651] env[62521]: DEBUG nova.compute.manager [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Stashing vm_state: active {{(pid=62521) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 835.517216] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318512, 'name': CloneVM_Task, 'duration_secs': 2.425182} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.517503] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Created linked-clone VM from snapshot [ 835.518478] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b6fe96-b3de-44cf-8859-e7ce2f1e3d60 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.529041] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Uploading image b2b9c4d6-2b30-4063-9fd3-400c421b1aac {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 835.546268] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 835.546558] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6be74fcf-8b03-4702-aeb8-0ef0c351d545 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.557308] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 835.557308] env[62521]: value = "task-1318520" [ 835.557308] env[62521]: _type = "Task" [ 835.557308] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.571020] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318520, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.598334] env[62521]: DEBUG nova.compute.manager [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.735087] env[62521]: DEBUG nova.network.neutron [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Successfully created port: 8c6d923d-1406-4e61-943b-a940584e6d6a {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.889104] env[62521]: DEBUG nova.network.neutron [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.946138] env[62521]: DEBUG nova.scheduler.client.report [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.959144] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318518, 'name': Rename_Task, 'duration_secs': 0.182918} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.959631] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 835.959946] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5bf567f-8b99-463a-b874-2823671bddc3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.967397] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 835.967397] env[62521]: value = "task-1318521" [ 835.967397] env[62521]: _type = "Task" [ 835.967397] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.984945] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318521, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.991649] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318519, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.016322] env[62521]: DEBUG nova.network.neutron [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Successfully created port: 2ffb5884-fc67-4969-b2e0-c5412152320f {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 836.026839] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.074163] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318520, 'name': Destroy_Task, 'duration_secs': 0.47847} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.074526] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Destroyed the VM [ 836.074865] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 836.076902] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1ceb8139-7d8d-47c5-9e4c-b07228cd5052 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.086117] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 836.086117] env[62521]: value = "task-1318522" [ 836.086117] env[62521]: _type = "Task" [ 836.086117] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.095235] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318522, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.097693] env[62521]: DEBUG nova.network.neutron [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.450035] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.863s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.452540] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.862s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.454310] env[62521]: INFO nova.compute.claims [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.473015] env[62521]: INFO nova.scheduler.client.report [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Deleted allocations for instance 50aef23e-9429-4499-9fc1-c31b97e23029 [ 836.484025] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318521, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.489715] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318519, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.689868} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.489715] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6/70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.490459] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.490459] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b313d9b-4808-46df-91b4-7c8927ff1a74 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.496917] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 836.496917] env[62521]: value = "task-1318523" [ 836.496917] env[62521]: _type = "Task" [ 836.496917] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.505412] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318523, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.596663] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318522, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.601371] env[62521]: DEBUG oslo_concurrency.lockutils [req-bef13e60-0105-40e1-b02c-b7f80bac099c req-74e62361-1a26-4316-9d67-64cd11b1ae7f service nova] Releasing lock "refresh_cache-4afe09db-6c01-444f-a127-6e1f97794544" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.601820] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquired lock "refresh_cache-4afe09db-6c01-444f-a127-6e1f97794544" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.602049] env[62521]: DEBUG nova.network.neutron [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 836.609865] env[62521]: DEBUG nova.compute.manager [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.636689] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.636932] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.637121] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.637309] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.637456] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.637603] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.637808] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.637966] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.638177] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.638349] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.638539] env[62521]: DEBUG nova.virt.hardware [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.639396] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d769ae98-0dcd-4f72-9228-bcd5e9dc380b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.647632] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813001da-2212-4e56-baab-1c2f9e715d35 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.979383] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318521, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.987799] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b402d8e1-0675-4a38-99da-73e33357f664 tempest-InstanceActionsV221TestJSON-1651300756 tempest-InstanceActionsV221TestJSON-1651300756-project-member] Lock "50aef23e-9429-4499-9fc1-c31b97e23029" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.334s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.010022] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318523, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117217} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.010022] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.010731] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed456aa6-f421-4838-8f17-d123bc052ffc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.032810] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6/70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.033115] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c36cc592-3815-41e4-8197-e10d5d97348e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.053478] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 837.053478] env[62521]: value = "task-1318524" [ 837.053478] env[62521]: _type = "Task" [ 837.053478] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.062031] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318524, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.097227] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318522, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.147166] env[62521]: DEBUG nova.network.neutron [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.479105] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318521, 'name': PowerOnVM_Task} progress is 68%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.484894] env[62521]: DEBUG nova.compute.manager [req-4c81a21a-02ca-42d4-889c-08f0ec106931 req-e2c3bee9-2bfc-4a05-bda8-a46a9618fb64 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received event network-vif-plugged-ef87a463-92f6-449b-8262-f156eaca1596 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.485122] env[62521]: DEBUG oslo_concurrency.lockutils [req-4c81a21a-02ca-42d4-889c-08f0ec106931 req-e2c3bee9-2bfc-4a05-bda8-a46a9618fb64 service nova] Acquiring lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.485331] env[62521]: DEBUG oslo_concurrency.lockutils [req-4c81a21a-02ca-42d4-889c-08f0ec106931 req-e2c3bee9-2bfc-4a05-bda8-a46a9618fb64 service nova] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.485492] env[62521]: DEBUG oslo_concurrency.lockutils [req-4c81a21a-02ca-42d4-889c-08f0ec106931 req-e2c3bee9-2bfc-4a05-bda8-a46a9618fb64 service nova] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.485656] env[62521]: DEBUG nova.compute.manager [req-4c81a21a-02ca-42d4-889c-08f0ec106931 req-e2c3bee9-2bfc-4a05-bda8-a46a9618fb64 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] No waiting events found dispatching network-vif-plugged-ef87a463-92f6-449b-8262-f156eaca1596 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 837.485895] env[62521]: WARNING nova.compute.manager [req-4c81a21a-02ca-42d4-889c-08f0ec106931 req-e2c3bee9-2bfc-4a05-bda8-a46a9618fb64 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received unexpected event network-vif-plugged-ef87a463-92f6-449b-8262-f156eaca1596 for instance with vm_state building and task_state spawning. [ 837.563576] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318524, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.579125] env[62521]: DEBUG nova.network.neutron [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Successfully updated port: ef87a463-92f6-449b-8262-f156eaca1596 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 837.601425] env[62521]: DEBUG oslo_vmware.api [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318522, 'name': RemoveSnapshot_Task, 'duration_secs': 1.12466} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.604167] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 837.682236] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ea2d09-ba93-4081-9916-8d1b7d57598e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.690350] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bd3735-c615-49eb-bd2a-5db05c8ea7b5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.725183] env[62521]: DEBUG nova.network.neutron [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Updating instance_info_cache with network_info: [{"id": "b520981a-8f32-4f58-9e84-a062b20d933a", "address": "fa:16:3e:da:1a:88", "network": {"id": "505063bb-bf4f-4b5d-8d34-c9baae0475a7", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-776289723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f87a25d6e31a47e397ebb9c6aa527305", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0d7a2b2f-3b49-4dc8-9096-af16144b27a9", "external-id": "nsx-vlan-transportzone-492", "segmentation_id": 492, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb520981a-8f", "ovs_interfaceid": "b520981a-8f32-4f58-9e84-a062b20d933a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.726614] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f636c519-fb5b-4f22-9027-e5aad3094441 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.735483] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ffdf20-f71b-4920-929b-25e2ac460f05 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.751567] env[62521]: DEBUG nova.compute.provider_tree [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.980560] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318521, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.068403] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318524, 'name': ReconfigVM_Task, 'duration_secs': 0.726969} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.068729] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6/70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.072911] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07c21214-c9df-49ca-b0a5-2173d7735202 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.076567] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 838.076567] env[62521]: value = "task-1318525" [ 838.076567] env[62521]: _type = "Task" [ 838.076567] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.093214] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318525, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.109894] env[62521]: WARNING nova.compute.manager [None req-3d204932-5cf7-4ddb-b847-d2310205227b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Image not found during snapshot: nova.exception.ImageNotFound: Image b2b9c4d6-2b30-4063-9fd3-400c421b1aac could not be found. [ 838.230577] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Releasing lock "refresh_cache-4afe09db-6c01-444f-a127-6e1f97794544" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.230928] env[62521]: DEBUG nova.compute.manager [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Instance network_info: |[{"id": "b520981a-8f32-4f58-9e84-a062b20d933a", "address": "fa:16:3e:da:1a:88", "network": {"id": "505063bb-bf4f-4b5d-8d34-c9baae0475a7", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-776289723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f87a25d6e31a47e397ebb9c6aa527305", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0d7a2b2f-3b49-4dc8-9096-af16144b27a9", "external-id": "nsx-vlan-transportzone-492", "segmentation_id": 492, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb520981a-8f", "ovs_interfaceid": "b520981a-8f32-4f58-9e84-a062b20d933a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 838.231366] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:1a:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0d7a2b2f-3b49-4dc8-9096-af16144b27a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b520981a-8f32-4f58-9e84-a062b20d933a', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 838.238627] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Creating folder: Project (f87a25d6e31a47e397ebb9c6aa527305). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.238883] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d67b21a1-6cef-4539-8be7-550d8442d397 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.250176] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Created folder: Project (f87a25d6e31a47e397ebb9c6aa527305) in parent group-v282025. [ 838.250351] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Creating folder: Instances. Parent ref: group-v282080. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.250566] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f38e7618-7f83-4680-a4d1-860cd2253675 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.253989] env[62521]: DEBUG nova.scheduler.client.report [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.261104] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Created folder: Instances in parent group-v282080. [ 838.261104] env[62521]: DEBUG oslo.service.loopingcall [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.261104] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 838.261104] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b0e7564-fed9-4729-94de-24d0cd3ef666 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.279667] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 838.279667] env[62521]: value = "task-1318528" [ 838.279667] env[62521]: _type = "Task" [ 838.279667] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.289802] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318528, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.482984] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318521, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.588767] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318525, 'name': Rename_Task, 'duration_secs': 0.124674} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.589926] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 838.590257] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b52e442-12d2-436c-88e1-e96f3b675fa9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.597401] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 838.597401] env[62521]: value = "task-1318529" [ 838.597401] env[62521]: _type = "Task" [ 838.597401] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.611089] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318529, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.758702] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.306s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.759271] env[62521]: DEBUG nova.compute.manager [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 838.762419] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.121s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.765924] env[62521]: INFO nova.compute.claims [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.791171] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318528, 'name': CreateVM_Task, 'duration_secs': 0.330779} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.791171] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.791664] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.791734] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.792088] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.792276] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bb2be93-725b-4d10-a006-56a73cb603bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.797212] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for the task: (returnval){ [ 838.797212] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b7004a-0fb6-cde0-c3b4-3a5b2ad134f3" [ 838.797212] env[62521]: _type = "Task" [ 838.797212] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.806919] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b7004a-0fb6-cde0-c3b4-3a5b2ad134f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.980488] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318521, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.110387] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318529, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.194118] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "b3e33bab-636f-42df-ac6c-667464312c11" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.194512] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "b3e33bab-636f-42df-ac6c-667464312c11" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.194840] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "b3e33bab-636f-42df-ac6c-667464312c11-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.194951] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "b3e33bab-636f-42df-ac6c-667464312c11-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.195318] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "b3e33bab-636f-42df-ac6c-667464312c11-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.197578] env[62521]: INFO nova.compute.manager [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Terminating instance [ 839.199543] env[62521]: DEBUG nova.compute.manager [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.199543] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 839.200368] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b583d3a-86b3-4682-b76f-d46a2ea6027e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.208878] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.208878] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1071c98-bc44-4347-8321-e72ad15f4591 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.214387] env[62521]: DEBUG oslo_vmware.api [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 839.214387] env[62521]: value = "task-1318530" [ 839.214387] env[62521]: _type = "Task" [ 839.214387] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.224132] env[62521]: DEBUG oslo_vmware.api [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318530, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.270207] env[62521]: DEBUG nova.compute.utils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.272775] env[62521]: DEBUG nova.compute.manager [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 839.272963] env[62521]: DEBUG nova.network.neutron [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 839.311899] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b7004a-0fb6-cde0-c3b4-3a5b2ad134f3, 'name': SearchDatastore_Task, 'duration_secs': 0.012466} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.312639] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.312907] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.313424] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.313762] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.314063] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.314481] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f830acb-1344-48e3-9442-e114a1ab4c95 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.320307] env[62521]: DEBUG nova.policy [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd3a17c5707746af8016a349024a7d3d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a9f1052b9944a76b87609da0aa6b8e6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.325122] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.325245] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.325943] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48c9a9d5-cbb8-40b2-9e2d-81c1faa28511 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.332428] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for the task: (returnval){ [ 839.332428] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e4120d-b9bf-396e-5e7e-057d5c597bac" [ 839.332428] env[62521]: _type = "Task" [ 839.332428] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.340895] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e4120d-b9bf-396e-5e7e-057d5c597bac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.481055] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318521, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.513633] env[62521]: DEBUG nova.compute.manager [req-5e291cfe-04a7-4dc7-80b3-eb50c4602686 req-5018cbcc-cb2c-4ccc-b1b6-8182d17034c6 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received event network-changed-ef87a463-92f6-449b-8262-f156eaca1596 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.513745] env[62521]: DEBUG nova.compute.manager [req-5e291cfe-04a7-4dc7-80b3-eb50c4602686 req-5018cbcc-cb2c-4ccc-b1b6-8182d17034c6 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Refreshing instance network info cache due to event network-changed-ef87a463-92f6-449b-8262-f156eaca1596. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 839.513940] env[62521]: DEBUG oslo_concurrency.lockutils [req-5e291cfe-04a7-4dc7-80b3-eb50c4602686 req-5018cbcc-cb2c-4ccc-b1b6-8182d17034c6 service nova] Acquiring lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.514099] env[62521]: DEBUG oslo_concurrency.lockutils [req-5e291cfe-04a7-4dc7-80b3-eb50c4602686 req-5018cbcc-cb2c-4ccc-b1b6-8182d17034c6 service nova] Acquired lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.514257] env[62521]: DEBUG nova.network.neutron [req-5e291cfe-04a7-4dc7-80b3-eb50c4602686 req-5018cbcc-cb2c-4ccc-b1b6-8182d17034c6 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Refreshing network info cache for port ef87a463-92f6-449b-8262-f156eaca1596 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 839.610656] env[62521]: DEBUG oslo_vmware.api [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318529, 'name': PowerOnVM_Task, 'duration_secs': 0.73439} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.611477] env[62521]: DEBUG nova.network.neutron [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Successfully created port: 892c26c7-9b46-4a0f-a406-6bdc9c0872fa {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.613501] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.613727] env[62521]: INFO nova.compute.manager [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Took 9.66 seconds to spawn the instance on the hypervisor. [ 839.613998] env[62521]: DEBUG nova.compute.manager [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.614839] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e508ea9-27af-41d4-99d3-6b5fb6763d20 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.725471] env[62521]: DEBUG oslo_vmware.api [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318530, 'name': PowerOffVM_Task, 'duration_secs': 0.208025} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.725757] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.725930] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 839.726207] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ade9a26-fd21-4597-a18f-90ea799845ee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.773220] env[62521]: DEBUG nova.compute.manager [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 839.796346] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 839.796346] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 839.796346] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Deleting the datastore file [datastore2] b3e33bab-636f-42df-ac6c-667464312c11 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.796346] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc823f66-26af-4cec-9f98-a16deda8b280 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.802171] env[62521]: DEBUG oslo_vmware.api [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 839.802171] env[62521]: value = "task-1318532" [ 839.802171] env[62521]: _type = "Task" [ 839.802171] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.810356] env[62521]: DEBUG oslo_vmware.api [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318532, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.849627] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e4120d-b9bf-396e-5e7e-057d5c597bac, 'name': SearchDatastore_Task, 'duration_secs': 0.010451} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.849992] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-948c29e1-7a8e-44d1-b42e-e9f044ecdaff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.855458] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for the task: (returnval){ [ 839.855458] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52473646-322a-daa7-232c-ddd011b34fbd" [ 839.855458] env[62521]: _type = "Task" [ 839.855458] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.868041] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52473646-322a-daa7-232c-ddd011b34fbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.920487] env[62521]: DEBUG nova.compute.manager [req-f6e3fce1-6b3b-44e3-9518-4bf4cbc4706d req-dfa479df-1b11-42ad-8a88-04eeedaa7814 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received event network-vif-plugged-8c6d923d-1406-4e61-943b-a940584e6d6a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.920772] env[62521]: DEBUG oslo_concurrency.lockutils [req-f6e3fce1-6b3b-44e3-9518-4bf4cbc4706d req-dfa479df-1b11-42ad-8a88-04eeedaa7814 service nova] Acquiring lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.921047] env[62521]: DEBUG oslo_concurrency.lockutils [req-f6e3fce1-6b3b-44e3-9518-4bf4cbc4706d req-dfa479df-1b11-42ad-8a88-04eeedaa7814 service nova] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.921347] env[62521]: DEBUG oslo_concurrency.lockutils [req-f6e3fce1-6b3b-44e3-9518-4bf4cbc4706d req-dfa479df-1b11-42ad-8a88-04eeedaa7814 service nova] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.921790] env[62521]: DEBUG nova.compute.manager [req-f6e3fce1-6b3b-44e3-9518-4bf4cbc4706d req-dfa479df-1b11-42ad-8a88-04eeedaa7814 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] No waiting events found dispatching network-vif-plugged-8c6d923d-1406-4e61-943b-a940584e6d6a {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.921790] env[62521]: WARNING nova.compute.manager [req-f6e3fce1-6b3b-44e3-9518-4bf4cbc4706d req-dfa479df-1b11-42ad-8a88-04eeedaa7814 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received unexpected event network-vif-plugged-8c6d923d-1406-4e61-943b-a940584e6d6a for instance with vm_state building and task_state spawning. [ 839.980690] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318521, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.042306] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e1ef9e-b5bf-4df8-8a9d-fc49fac3d920 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.050737] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ba192b-da45-4729-9d3d-f47b90f95c80 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.087891] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0819d48-60d6-4fd0-b6da-91af1f77e8ea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.096208] env[62521]: DEBUG nova.network.neutron [req-5e291cfe-04a7-4dc7-80b3-eb50c4602686 req-5018cbcc-cb2c-4ccc-b1b6-8182d17034c6 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.100717] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc77ab65-a916-473f-b347-f813d5f0b852 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.118181] env[62521]: DEBUG nova.compute.provider_tree [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.130932] env[62521]: DEBUG nova.network.neutron [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Successfully updated port: 8c6d923d-1406-4e61-943b-a940584e6d6a {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.134692] env[62521]: INFO nova.compute.manager [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Took 41.03 seconds to build instance. [ 840.314465] env[62521]: DEBUG oslo_vmware.api [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318532, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224139} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.315015] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.315164] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 840.315399] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 840.315646] env[62521]: INFO nova.compute.manager [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Took 1.12 seconds to destroy the instance on the hypervisor. [ 840.316079] env[62521]: DEBUG oslo.service.loopingcall [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.316303] env[62521]: DEBUG nova.compute.manager [-] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.316402] env[62521]: DEBUG nova.network.neutron [-] [instance: b3e33bab-636f-42df-ac6c-667464312c11] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 840.365838] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52473646-322a-daa7-232c-ddd011b34fbd, 'name': SearchDatastore_Task, 'duration_secs': 0.01111} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.366841] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.367130] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 4afe09db-6c01-444f-a127-6e1f97794544/4afe09db-6c01-444f-a127-6e1f97794544.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 840.367391] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ddd07ae9-18b2-457e-a35e-04824fcb1d5b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.374294] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for the task: (returnval){ [ 840.374294] env[62521]: value = "task-1318533" [ 840.374294] env[62521]: _type = "Task" [ 840.374294] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.382560] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.453936] env[62521]: DEBUG nova.network.neutron [req-5e291cfe-04a7-4dc7-80b3-eb50c4602686 req-5018cbcc-cb2c-4ccc-b1b6-8182d17034c6 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.481918] env[62521]: DEBUG oslo_vmware.api [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318521, 'name': PowerOnVM_Task, 'duration_secs': 4.392913} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.482156] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.482367] env[62521]: INFO nova.compute.manager [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Took 12.93 seconds to spawn the instance on the hypervisor. [ 840.482544] env[62521]: DEBUG nova.compute.manager [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.483333] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d24041-7353-461a-b978-7bf8a2eada3c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.621317] env[62521]: DEBUG nova.scheduler.client.report [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.639270] env[62521]: DEBUG oslo_concurrency.lockutils [None req-920c9d93-4416-4409-a7ad-7eee4f550a58 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.809s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.787339] env[62521]: DEBUG nova.compute.manager [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 840.836036] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.836300] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.836460] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.836629] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.836773] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.836971] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.837286] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.837437] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.837601] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.837760] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.837930] env[62521]: DEBUG nova.virt.hardware [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.838873] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744193e1-6f51-4380-97d1-19be32d05cd0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.847415] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abf48801-e5ab-4691-b8f0-3d0973a80acd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.892054] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318533, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.957099] env[62521]: DEBUG oslo_concurrency.lockutils [req-5e291cfe-04a7-4dc7-80b3-eb50c4602686 req-5018cbcc-cb2c-4ccc-b1b6-8182d17034c6 service nova] Releasing lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.003013] env[62521]: INFO nova.compute.manager [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Took 43.82 seconds to build instance. [ 841.126009] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.126562] env[62521]: DEBUG nova.compute.manager [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 841.130200] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.039s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.132427] env[62521]: INFO nova.compute.claims [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.196020] env[62521]: DEBUG nova.network.neutron [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Successfully updated port: 892c26c7-9b46-4a0f-a406-6bdc9c0872fa {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.266805] env[62521]: DEBUG nova.network.neutron [-] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.391688] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318533, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575204} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.391961] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 4afe09db-6c01-444f-a127-6e1f97794544/4afe09db-6c01-444f-a127-6e1f97794544.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 841.392191] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 841.392710] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2bacf32-9278-4082-bd41-c948e2a72364 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.400244] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for the task: (returnval){ [ 841.400244] env[62521]: value = "task-1318534" [ 841.400244] env[62521]: _type = "Task" [ 841.400244] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.408482] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318534, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.505619] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ee04c9a5-aa5e-459c-8f8a-06397e139dd3 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "4dcd1c6c-8726-42db-997a-e78ba1293310" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.433s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.637873] env[62521]: DEBUG nova.compute.utils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.644262] env[62521]: DEBUG nova.compute.manager [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.644427] env[62521]: DEBUG nova.network.neutron [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 841.650670] env[62521]: DEBUG nova.compute.manager [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Received event network-vif-plugged-892c26c7-9b46-4a0f-a406-6bdc9c0872fa {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.651341] env[62521]: DEBUG oslo_concurrency.lockutils [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] Acquiring lock "92629ebf-eb50-4862-abe2-01e210e302ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.651577] env[62521]: DEBUG oslo_concurrency.lockutils [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] Lock "92629ebf-eb50-4862-abe2-01e210e302ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.652096] env[62521]: DEBUG oslo_concurrency.lockutils [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] Lock "92629ebf-eb50-4862-abe2-01e210e302ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.652096] env[62521]: DEBUG nova.compute.manager [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] No waiting events found dispatching network-vif-plugged-892c26c7-9b46-4a0f-a406-6bdc9c0872fa {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.652096] env[62521]: WARNING nova.compute.manager [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Received unexpected event network-vif-plugged-892c26c7-9b46-4a0f-a406-6bdc9c0872fa for instance with vm_state building and task_state spawning. [ 841.652244] env[62521]: DEBUG nova.compute.manager [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Received event network-changed-892c26c7-9b46-4a0f-a406-6bdc9c0872fa {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.652365] env[62521]: DEBUG nova.compute.manager [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Refreshing instance network info cache due to event network-changed-892c26c7-9b46-4a0f-a406-6bdc9c0872fa. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 841.652544] env[62521]: DEBUG oslo_concurrency.lockutils [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] Acquiring lock "refresh_cache-92629ebf-eb50-4862-abe2-01e210e302ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.652666] env[62521]: DEBUG oslo_concurrency.lockutils [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] Acquired lock "refresh_cache-92629ebf-eb50-4862-abe2-01e210e302ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.652813] env[62521]: DEBUG nova.network.neutron [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Refreshing network info cache for port 892c26c7-9b46-4a0f-a406-6bdc9c0872fa {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 841.698477] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquiring lock "refresh_cache-92629ebf-eb50-4862-abe2-01e210e302ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.700370] env[62521]: DEBUG nova.policy [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f044b14654f45a9b40143285f4b1672', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a592d0f3c5834858bda6ca89a33662a9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.769635] env[62521]: INFO nova.compute.manager [-] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Took 1.45 seconds to deallocate network for instance. [ 841.911579] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318534, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123466} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.911904] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.912814] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92892a4e-fd4d-4952-8b32-369fa91bc42c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.942512] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 4afe09db-6c01-444f-a127-6e1f97794544/4afe09db-6c01-444f-a127-6e1f97794544.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.943545] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a31b19c-76cc-42aa-a930-933d63db7465 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.964391] env[62521]: DEBUG nova.compute.manager [req-721e331c-54aa-48fd-93a2-9026d22d93cf req-59185a5c-f4eb-4349-82b4-d68629e5fcfa service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received event network-changed-8c6d923d-1406-4e61-943b-a940584e6d6a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.964518] env[62521]: DEBUG nova.compute.manager [req-721e331c-54aa-48fd-93a2-9026d22d93cf req-59185a5c-f4eb-4349-82b4-d68629e5fcfa service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Refreshing instance network info cache due to event network-changed-8c6d923d-1406-4e61-943b-a940584e6d6a. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 841.964738] env[62521]: DEBUG oslo_concurrency.lockutils [req-721e331c-54aa-48fd-93a2-9026d22d93cf req-59185a5c-f4eb-4349-82b4-d68629e5fcfa service nova] Acquiring lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.964879] env[62521]: DEBUG oslo_concurrency.lockutils [req-721e331c-54aa-48fd-93a2-9026d22d93cf req-59185a5c-f4eb-4349-82b4-d68629e5fcfa service nova] Acquired lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.965051] env[62521]: DEBUG nova.network.neutron [req-721e331c-54aa-48fd-93a2-9026d22d93cf req-59185a5c-f4eb-4349-82b4-d68629e5fcfa service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Refreshing network info cache for port 8c6d923d-1406-4e61-943b-a940584e6d6a {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 841.969546] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for the task: (returnval){ [ 841.969546] env[62521]: value = "task-1318535" [ 841.969546] env[62521]: _type = "Task" [ 841.969546] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.980344] env[62521]: INFO nova.compute.manager [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Rebuilding instance [ 841.982272] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318535, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.990592] env[62521]: DEBUG nova.network.neutron [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Successfully created port: 903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.023850] env[62521]: DEBUG nova.compute.manager [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.024857] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0be742-4d1f-419b-855f-ded9c2cc5ffa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.145439] env[62521]: DEBUG nova.compute.manager [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 842.223255] env[62521]: DEBUG nova.network.neutron [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.275480] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.361203] env[62521]: DEBUG nova.network.neutron [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.391086] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c81f093-eaca-475a-824a-12b97d34b6ab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.399090] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0169f71e-19a5-469a-b6c4-75cd26fb232b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.432978] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba65d72b-27ea-4b7d-8568-1d6be8ebfd4e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.442346] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f0209f-b934-4ea1-95d6-d8311cd4ea5b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.456627] env[62521]: DEBUG nova.compute.provider_tree [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.480550] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318535, 'name': ReconfigVM_Task, 'duration_secs': 0.291894} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.480974] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 4afe09db-6c01-444f-a127-6e1f97794544/4afe09db-6c01-444f-a127-6e1f97794544.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.481854] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9a3737f2-bb47-4bd3-9b65-fcadd365ba8f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.488534] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for the task: (returnval){ [ 842.488534] env[62521]: value = "task-1318536" [ 842.488534] env[62521]: _type = "Task" [ 842.488534] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.497327] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318536, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.539639] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 842.542049] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44e66a73-ffda-4e70-9726-d98ad8b29482 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.547243] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 842.547243] env[62521]: value = "task-1318537" [ 842.547243] env[62521]: _type = "Task" [ 842.547243] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.556827] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.728627] env[62521]: DEBUG nova.network.neutron [req-721e331c-54aa-48fd-93a2-9026d22d93cf req-59185a5c-f4eb-4349-82b4-d68629e5fcfa service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.867128] env[62521]: DEBUG oslo_concurrency.lockutils [req-ae1f2f87-859e-4601-bb7f-e71ff92cc892 req-326ad11e-d1b0-47b2-8e6b-6f98d3a7dc41 service nova] Releasing lock "refresh_cache-92629ebf-eb50-4862-abe2-01e210e302ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.867629] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquired lock "refresh_cache-92629ebf-eb50-4862-abe2-01e210e302ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.867914] env[62521]: DEBUG nova.network.neutron [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.900442] env[62521]: DEBUG nova.network.neutron [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Successfully updated port: 2ffb5884-fc67-4969-b2e0-c5412152320f {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.944315] env[62521]: DEBUG nova.network.neutron [req-721e331c-54aa-48fd-93a2-9026d22d93cf req-59185a5c-f4eb-4349-82b4-d68629e5fcfa service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.959728] env[62521]: DEBUG nova.scheduler.client.report [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.973518] env[62521]: DEBUG nova.compute.manager [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.974469] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ad51a3-ba3e-42a3-8d51-31d436920da5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.000360] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318536, 'name': Rename_Task, 'duration_secs': 0.154627} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.000908] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 843.001339] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35c15b3e-0388-453a-8710-4ecde5968353 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.009881] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for the task: (returnval){ [ 843.009881] env[62521]: value = "task-1318538" [ 843.009881] env[62521]: _type = "Task" [ 843.009881] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.022576] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318538, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.060203] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318537, 'name': PowerOffVM_Task, 'duration_secs': 0.424292} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.060592] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.060746] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 843.061578] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abf91779-9aa8-4bfd-88cb-861eb817c9c6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.069395] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.069671] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8897a79e-e6f6-4d70-bb57-218464aff6af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.154088] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.154405] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.154405] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleting the datastore file [datastore1] 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.154710] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55dee2da-4501-4303-9579-58222e8a0d2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.158050] env[62521]: DEBUG nova.compute.manager [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 843.167209] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 843.167209] env[62521]: value = "task-1318540" [ 843.167209] env[62521]: _type = "Task" [ 843.167209] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.174992] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318540, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.183134] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.183412] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.183582] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.184095] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.184095] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.184095] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.184338] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.184548] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.184702] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.184879] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.185079] env[62521]: DEBUG nova.virt.hardware [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.185984] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d19562d-a86e-42c2-aaa9-910b5185658b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.194248] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50091540-42e4-4255-8f23-580380c93620 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.402616] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.436532] env[62521]: DEBUG nova.network.neutron [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.448318] env[62521]: DEBUG oslo_concurrency.lockutils [req-721e331c-54aa-48fd-93a2-9026d22d93cf req-59185a5c-f4eb-4349-82b4-d68629e5fcfa service nova] Releasing lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.448560] env[62521]: DEBUG nova.compute.manager [req-721e331c-54aa-48fd-93a2-9026d22d93cf req-59185a5c-f4eb-4349-82b4-d68629e5fcfa service nova] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Received event network-vif-deleted-a2a1e281-79b9-4cf2-9229-c2e6acb45ec7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.448889] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquired lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.449049] env[62521]: DEBUG nova.network.neutron [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.464057] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.464625] env[62521]: DEBUG nova.compute.manager [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 843.468643] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.516s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.471080] env[62521]: INFO nova.compute.claims [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.492209] env[62521]: INFO nova.compute.manager [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] instance snapshotting [ 843.494755] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b213c77c-6f36-4b4a-a677-9d35ac189f74 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.521730] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409c7e96-51b0-48c7-93df-7f700059d983 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.531202] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318538, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.625159] env[62521]: DEBUG nova.network.neutron [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Updating instance_info_cache with network_info: [{"id": "892c26c7-9b46-4a0f-a406-6bdc9c0872fa", "address": "fa:16:3e:bd:07:7d", "network": {"id": "a013b7aa-023d-4fc6-b0af-978a0806e79f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1099094943-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a9f1052b9944a76b87609da0aa6b8e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap892c26c7-9b", "ovs_interfaceid": "892c26c7-9b46-4a0f-a406-6bdc9c0872fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.683755] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318540, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154815} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.683755] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.683755] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 843.683755] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 843.683755] env[62521]: DEBUG nova.compute.manager [req-8d374615-8bcc-486d-8a57-cb2c8b2bd578 req-8171ed44-d820-46c6-a067-296574e1f0c2 service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Received event network-vif-plugged-903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.683898] env[62521]: DEBUG oslo_concurrency.lockutils [req-8d374615-8bcc-486d-8a57-cb2c8b2bd578 req-8171ed44-d820-46c6-a067-296574e1f0c2 service nova] Acquiring lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.683898] env[62521]: DEBUG oslo_concurrency.lockutils [req-8d374615-8bcc-486d-8a57-cb2c8b2bd578 req-8171ed44-d820-46c6-a067-296574e1f0c2 service nova] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.683898] env[62521]: DEBUG oslo_concurrency.lockutils [req-8d374615-8bcc-486d-8a57-cb2c8b2bd578 req-8171ed44-d820-46c6-a067-296574e1f0c2 service nova] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.683898] env[62521]: DEBUG nova.compute.manager [req-8d374615-8bcc-486d-8a57-cb2c8b2bd578 req-8171ed44-d820-46c6-a067-296574e1f0c2 service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] No waiting events found dispatching network-vif-plugged-903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 843.683898] env[62521]: WARNING nova.compute.manager [req-8d374615-8bcc-486d-8a57-cb2c8b2bd578 req-8171ed44-d820-46c6-a067-296574e1f0c2 service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Received unexpected event network-vif-plugged-903b74d0-11d2-404d-81f1-b6bb18f713af for instance with vm_state building and task_state spawning. [ 843.684043] env[62521]: DEBUG nova.network.neutron [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Successfully updated port: 903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.969816] env[62521]: DEBUG nova.compute.utils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.971481] env[62521]: DEBUG nova.compute.manager [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.971481] env[62521]: DEBUG nova.network.neutron [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.991448] env[62521]: DEBUG nova.network.neutron [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.024438] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318538, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.037379] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 844.037783] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-838a8a92-dd5a-462c-9494-d7ac730c37a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.044636] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 844.044636] env[62521]: value = "task-1318541" [ 844.044636] env[62521]: _type = "Task" [ 844.044636] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.053570] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318541, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.075240] env[62521]: DEBUG nova.policy [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e602eba085748d59dc453e05c98e39e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ab001785c25414abbaa4bfe3a6a0bb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 844.126245] env[62521]: DEBUG nova.compute.manager [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received event network-vif-plugged-2ffb5884-fc67-4969-b2e0-c5412152320f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.126245] env[62521]: DEBUG oslo_concurrency.lockutils [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] Acquiring lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.126245] env[62521]: DEBUG oslo_concurrency.lockutils [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.126245] env[62521]: DEBUG oslo_concurrency.lockutils [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.126245] env[62521]: DEBUG nova.compute.manager [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] No waiting events found dispatching network-vif-plugged-2ffb5884-fc67-4969-b2e0-c5412152320f {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.126443] env[62521]: WARNING nova.compute.manager [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received unexpected event network-vif-plugged-2ffb5884-fc67-4969-b2e0-c5412152320f for instance with vm_state building and task_state spawning. [ 844.127415] env[62521]: DEBUG nova.compute.manager [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received event network-changed-2ffb5884-fc67-4969-b2e0-c5412152320f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.127771] env[62521]: DEBUG nova.compute.manager [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Refreshing instance network info cache due to event network-changed-2ffb5884-fc67-4969-b2e0-c5412152320f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.128621] env[62521]: DEBUG oslo_concurrency.lockutils [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] Acquiring lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.129824] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Releasing lock "refresh_cache-92629ebf-eb50-4862-abe2-01e210e302ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.130344] env[62521]: DEBUG nova.compute.manager [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Instance network_info: |[{"id": "892c26c7-9b46-4a0f-a406-6bdc9c0872fa", "address": "fa:16:3e:bd:07:7d", "network": {"id": "a013b7aa-023d-4fc6-b0af-978a0806e79f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1099094943-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a9f1052b9944a76b87609da0aa6b8e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap892c26c7-9b", "ovs_interfaceid": "892c26c7-9b46-4a0f-a406-6bdc9c0872fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.131095] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:07:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73f6629b-7f80-4a5b-8f15-c7a1635b3c33', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '892c26c7-9b46-4a0f-a406-6bdc9c0872fa', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.140241] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Creating folder: Project (7a9f1052b9944a76b87609da0aa6b8e6). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.143398] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a334107a-b6ba-4025-8e94-81e5169f61f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.154740] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Created folder: Project (7a9f1052b9944a76b87609da0aa6b8e6) in parent group-v282025. [ 844.155675] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Creating folder: Instances. Parent ref: group-v282083. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.156106] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44062b8c-a9bf-4d3b-9f60-2c84745ec669 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.165731] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Created folder: Instances in parent group-v282083. [ 844.166360] env[62521]: DEBUG oslo.service.loopingcall [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.166655] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.167036] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55400f7b-9416-4bdf-b699-84049bf377d9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.187107] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.187107] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.187107] env[62521]: DEBUG nova.network.neutron [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.197655] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.197655] env[62521]: value = "task-1318544" [ 844.197655] env[62521]: _type = "Task" [ 844.197655] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.207944] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318544, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.467514] env[62521]: DEBUG nova.network.neutron [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Updating instance_info_cache with network_info: [{"id": "ef87a463-92f6-449b-8262-f156eaca1596", "address": "fa:16:3e:d7:cc:7e", "network": {"id": "41bc38b6-ff85-44c9-b5b9-1624a2a3368c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2055921925", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.183", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef87a463-92", "ovs_interfaceid": "ef87a463-92f6-449b-8262-f156eaca1596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c6d923d-1406-4e61-943b-a940584e6d6a", "address": "fa:16:3e:82:86:71", "network": {"id": "8218260a-d7e4-469a-bfb9-d5951e9a7e56", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1300759567", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46785c9c-8b22-487d-a854-b3e67c5ed1d7", "external-id": "nsx-vlan-transportzone-430", "segmentation_id": 430, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c6d923d-14", "ovs_interfaceid": "8c6d923d-1406-4e61-943b-a940584e6d6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2ffb5884-fc67-4969-b2e0-c5412152320f", "address": "fa:16:3e:88:95:18", "network": {"id": "41bc38b6-ff85-44c9-b5b9-1624a2a3368c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2055921925", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.166", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ffb5884-fc", "ovs_interfaceid": "2ffb5884-fc67-4969-b2e0-c5412152320f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.475343] env[62521]: DEBUG nova.compute.manager [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 844.530037] env[62521]: DEBUG oslo_vmware.api [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318538, 'name': PowerOnVM_Task, 'duration_secs': 1.51746} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.530662] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 844.530735] env[62521]: INFO nova.compute.manager [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Took 10.40 seconds to spawn the instance on the hypervisor. [ 844.530908] env[62521]: DEBUG nova.compute.manager [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.532025] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11551662-1ebf-4698-911e-9356c9de4651 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.557214] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318541, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.618997] env[62521]: DEBUG nova.network.neutron [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Successfully created port: 72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.713020] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318544, 'name': CreateVM_Task, 'duration_secs': 0.336546} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.713457] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 844.714701] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.716041] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.716041] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.719135] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31f656ea-192b-4c32-88f5-c229d883ddd6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.727793] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 844.727793] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e0c22f-ef99-a16a-f1a3-e1cfc17262be" [ 844.727793] env[62521]: _type = "Task" [ 844.727793] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.734024] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.734567] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.736376] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.736376] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.736376] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.736376] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.736376] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.736376] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.736376] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.736376] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.736376] env[62521]: DEBUG nova.virt.hardware [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.737607] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cbe5ec-88fd-4784-b075-4fa567dfa3d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.740429] env[62521]: DEBUG nova.network.neutron [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.750426] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e0c22f-ef99-a16a-f1a3-e1cfc17262be, 'name': SearchDatastore_Task, 'duration_secs': 0.009725} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.752437] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.753170] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.753170] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.753170] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.753311] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.753481] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b81c02fe-2705-44fd-a12a-a5d26f83c6de {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.756588] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0abc82d1-d6f5-4c74-85d2-1baf8bb44d2d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.776092] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:86:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7d4852b-d44e-448f-b25d-d4477b2c3ed9', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.784470] env[62521]: DEBUG oslo.service.loopingcall [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.787212] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.787831] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.788017] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.789126] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35f3955-44e8-48d7-9d43-e415f8ed5919 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.794013] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd960427-a94d-404b-9d4c-6fdaf3026971 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.809066] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0f37650-1a5f-41f1-8ccf-fc6eb418836c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.819493] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d917eb1e-e90a-42bf-bea8-e0183bb69d57 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.822276] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.822276] env[62521]: value = "task-1318545" [ 844.822276] env[62521]: _type = "Task" [ 844.822276] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.822566] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 844.822566] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b79be4-04ec-d7de-3b36-e56d2fabbe2f" [ 844.822566] env[62521]: _type = "Task" [ 844.822566] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.859878] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc35c9b0-08dc-45fd-9848-5e4df5083329 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.869419] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318545, 'name': CreateVM_Task} progress is 15%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.869750] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b79be4-04ec-d7de-3b36-e56d2fabbe2f, 'name': SearchDatastore_Task, 'duration_secs': 0.009022} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.872324] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec028d78-4c7a-4236-a640-fbcfa99ff6be {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.877470] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1442cc94-59df-49da-a291-3395e4cf5c1b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.882988] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 844.882988] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b5f85d-a68b-4728-666f-7cb483f5e4d9" [ 844.882988] env[62521]: _type = "Task" [ 844.882988] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.894463] env[62521]: DEBUG nova.compute.provider_tree [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.901982] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b5f85d-a68b-4728-666f-7cb483f5e4d9, 'name': SearchDatastore_Task, 'duration_secs': 0.015159} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.901982] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.901982] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 92629ebf-eb50-4862-abe2-01e210e302ad/92629ebf-eb50-4862-abe2-01e210e302ad.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.902289] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d70e726-a439-480c-90a2-bf96e26e4dde {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.908936] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 844.908936] env[62521]: value = "task-1318546" [ 844.908936] env[62521]: _type = "Task" [ 844.908936] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.919525] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318546, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.961711] env[62521]: DEBUG nova.network.neutron [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updating instance_info_cache with network_info: [{"id": "903b74d0-11d2-404d-81f1-b6bb18f713af", "address": "fa:16:3e:6a:82:e1", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap903b74d0-11", "ovs_interfaceid": "903b74d0-11d2-404d-81f1-b6bb18f713af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.971313] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Releasing lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.971313] env[62521]: DEBUG nova.compute.manager [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Instance network_info: |[{"id": "ef87a463-92f6-449b-8262-f156eaca1596", "address": "fa:16:3e:d7:cc:7e", "network": {"id": "41bc38b6-ff85-44c9-b5b9-1624a2a3368c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2055921925", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.183", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef87a463-92", "ovs_interfaceid": "ef87a463-92f6-449b-8262-f156eaca1596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c6d923d-1406-4e61-943b-a940584e6d6a", "address": "fa:16:3e:82:86:71", "network": {"id": "8218260a-d7e4-469a-bfb9-d5951e9a7e56", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1300759567", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46785c9c-8b22-487d-a854-b3e67c5ed1d7", "external-id": "nsx-vlan-transportzone-430", "segmentation_id": 430, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c6d923d-14", "ovs_interfaceid": "8c6d923d-1406-4e61-943b-a940584e6d6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2ffb5884-fc67-4969-b2e0-c5412152320f", "address": "fa:16:3e:88:95:18", "network": {"id": "41bc38b6-ff85-44c9-b5b9-1624a2a3368c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2055921925", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.166", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ffb5884-fc", "ovs_interfaceid": "2ffb5884-fc67-4969-b2e0-c5412152320f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.971504] env[62521]: DEBUG oslo_concurrency.lockutils [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] Acquired lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.971708] env[62521]: DEBUG nova.network.neutron [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Refreshing network info cache for port 2ffb5884-fc67-4969-b2e0-c5412152320f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.974209] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:cc:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef87a463-92f6-449b-8262-f156eaca1596', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:86:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '46785c9c-8b22-487d-a854-b3e67c5ed1d7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c6d923d-1406-4e61-943b-a940584e6d6a', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:95:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ffb5884-fc67-4969-b2e0-c5412152320f', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.983996] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Creating folder: Project (4facd01ddac841298d07810781c61e05). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.988840] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb1313e5-f954-4c8a-8321-6fde0015def7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.999249] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Created folder: Project (4facd01ddac841298d07810781c61e05) in parent group-v282025. [ 844.999477] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Creating folder: Instances. Parent ref: group-v282088. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.999751] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e331918-ae05-44db-9f82-a16d44f8b75f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.009163] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Created folder: Instances in parent group-v282088. [ 845.009398] env[62521]: DEBUG oslo.service.loopingcall [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.009588] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.009795] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b31c6281-2597-44c6-80e4-c6af24a77c5e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.034501] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.034501] env[62521]: value = "task-1318549" [ 845.034501] env[62521]: _type = "Task" [ 845.034501] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.042295] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318549, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.054648] env[62521]: INFO nova.compute.manager [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Took 31.88 seconds to build instance. [ 845.059010] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318541, 'name': CreateSnapshot_Task, 'duration_secs': 0.823644} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.059323] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 845.060663] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f45255e-6005-46a0-b4c4-70daa736b4e8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.336283] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318545, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.397864] env[62521]: DEBUG nova.scheduler.client.report [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.421842] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318546, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.464188] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.464695] env[62521]: DEBUG nova.compute.manager [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Instance network_info: |[{"id": "903b74d0-11d2-404d-81f1-b6bb18f713af", "address": "fa:16:3e:6a:82:e1", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap903b74d0-11", "ovs_interfaceid": "903b74d0-11d2-404d-81f1-b6bb18f713af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.464971] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:82:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd646f9d5-d2ad-4c22-bea5-85a965334de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '903b74d0-11d2-404d-81f1-b6bb18f713af', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.473130] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Creating folder: Project (a592d0f3c5834858bda6ca89a33662a9). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.473346] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52f17035-70f5-4782-af30-e3850efd9656 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.489482] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Created folder: Project (a592d0f3c5834858bda6ca89a33662a9) in parent group-v282025. [ 845.489707] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Creating folder: Instances. Parent ref: group-v282091. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.490318] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-38071d10-ed7d-41ff-997f-e844e711ae2e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.492677] env[62521]: DEBUG nova.compute.manager [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 845.502666] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Created folder: Instances in parent group-v282091. [ 845.503404] env[62521]: DEBUG oslo.service.loopingcall [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.503404] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.503584] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d52dd964-6dd4-4613-90cc-a19e3f160ee9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.525451] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.525451] env[62521]: value = "task-1318552" [ 845.525451] env[62521]: _type = "Task" [ 845.525451] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.529817] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.529980] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.530177] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.530344] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.530493] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.530653] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.530921] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.531042] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.531202] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.531363] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.531531] env[62521]: DEBUG nova.virt.hardware [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.534541] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5206085b-847e-49fc-9c5d-de98c8f2b2b8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.548581] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318552, 'name': CreateVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.552888] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81188ecb-1f35-4238-aa9f-121ccaaf7e39 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.565025] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8c93ed2f-3ccb-4c2b-be64-4e2741f78d34 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lock "4afe09db-6c01-444f-a127-6e1f97794544" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.019s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.565025] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318549, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.580494] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 845.583579] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9f2f9b6e-f06b-411b-ab92-ccee711677cf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.592478] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 845.592478] env[62521]: value = "task-1318553" [ 845.592478] env[62521]: _type = "Task" [ 845.592478] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.601326] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318553, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.718847] env[62521]: DEBUG nova.compute.manager [req-a7ffb8de-ea30-4a92-9286-2765eb5cf4be req-844c2bff-6441-46ec-b159-209fd8fe3a8b service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Received event network-changed-903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.719069] env[62521]: DEBUG nova.compute.manager [req-a7ffb8de-ea30-4a92-9286-2765eb5cf4be req-844c2bff-6441-46ec-b159-209fd8fe3a8b service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Refreshing instance network info cache due to event network-changed-903b74d0-11d2-404d-81f1-b6bb18f713af. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.719401] env[62521]: DEBUG oslo_concurrency.lockutils [req-a7ffb8de-ea30-4a92-9286-2765eb5cf4be req-844c2bff-6441-46ec-b159-209fd8fe3a8b service nova] Acquiring lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.719510] env[62521]: DEBUG oslo_concurrency.lockutils [req-a7ffb8de-ea30-4a92-9286-2765eb5cf4be req-844c2bff-6441-46ec-b159-209fd8fe3a8b service nova] Acquired lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.719676] env[62521]: DEBUG nova.network.neutron [req-a7ffb8de-ea30-4a92-9286-2765eb5cf4be req-844c2bff-6441-46ec-b159-209fd8fe3a8b service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Refreshing network info cache for port 903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 845.754169] env[62521]: DEBUG nova.network.neutron [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Updated VIF entry in instance network info cache for port 2ffb5884-fc67-4969-b2e0-c5412152320f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.754169] env[62521]: DEBUG nova.network.neutron [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Updating instance_info_cache with network_info: [{"id": "ef87a463-92f6-449b-8262-f156eaca1596", "address": "fa:16:3e:d7:cc:7e", "network": {"id": "41bc38b6-ff85-44c9-b5b9-1624a2a3368c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2055921925", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.183", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef87a463-92", "ovs_interfaceid": "ef87a463-92f6-449b-8262-f156eaca1596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c6d923d-1406-4e61-943b-a940584e6d6a", "address": "fa:16:3e:82:86:71", "network": {"id": "8218260a-d7e4-469a-bfb9-d5951e9a7e56", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1300759567", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46785c9c-8b22-487d-a854-b3e67c5ed1d7", "external-id": "nsx-vlan-transportzone-430", "segmentation_id": 430, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c6d923d-14", "ovs_interfaceid": "8c6d923d-1406-4e61-943b-a940584e6d6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2ffb5884-fc67-4969-b2e0-c5412152320f", "address": "fa:16:3e:88:95:18", "network": {"id": "41bc38b6-ff85-44c9-b5b9-1624a2a3368c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2055921925", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.166", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ffb5884-fc", "ovs_interfaceid": "2ffb5884-fc67-4969-b2e0-c5412152320f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.835412] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318545, 'name': CreateVM_Task, 'duration_secs': 0.598105} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.835565] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.836716] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.836716] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.837025] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.838206] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f645168a-3a9b-4dce-b356-2f0f6d06f1ea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.842446] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 845.842446] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526fcf0a-0b67-8e5d-3d20-2532432a85b2" [ 845.842446] env[62521]: _type = "Task" [ 845.842446] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.851341] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526fcf0a-0b67-8e5d-3d20-2532432a85b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.903281] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.904071] env[62521]: DEBUG nova.compute.manager [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.906614] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.929s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.908239] env[62521]: INFO nova.compute.claims [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.921701] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318546, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512056} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.921701] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 92629ebf-eb50-4862-abe2-01e210e302ad/92629ebf-eb50-4862-abe2-01e210e302ad.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.921701] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.921701] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-489977e8-8256-483e-9f72-fd3e4d18765e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.929322] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 845.929322] env[62521]: value = "task-1318554" [ 845.929322] env[62521]: _type = "Task" [ 845.929322] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.940415] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318554, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.036674] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318552, 'name': CreateVM_Task, 'duration_secs': 0.361126} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.037463] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.038529] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.038844] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.039300] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.044231] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19be22a2-a9cd-40a3-9c7b-2ccb8f1aafb7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.050102] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 846.050102] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cc5904-643f-e61a-6e06-b801994aeb66" [ 846.050102] env[62521]: _type = "Task" [ 846.050102] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.056401] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318549, 'name': CreateVM_Task, 'duration_secs': 0.739921} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.059922] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.060921] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.067991] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cc5904-643f-e61a-6e06-b801994aeb66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.105784] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318553, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.195722] env[62521]: DEBUG nova.compute.manager [req-7bec6aee-d58c-4597-a802-50d794de292d req-f5bae662-b21f-4a43-8502-7c917c26c897 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received event network-vif-plugged-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.196039] env[62521]: DEBUG oslo_concurrency.lockutils [req-7bec6aee-d58c-4597-a802-50d794de292d req-f5bae662-b21f-4a43-8502-7c917c26c897 service nova] Acquiring lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.196262] env[62521]: DEBUG oslo_concurrency.lockutils [req-7bec6aee-d58c-4597-a802-50d794de292d req-f5bae662-b21f-4a43-8502-7c917c26c897 service nova] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.196434] env[62521]: DEBUG oslo_concurrency.lockutils [req-7bec6aee-d58c-4597-a802-50d794de292d req-f5bae662-b21f-4a43-8502-7c917c26c897 service nova] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.196605] env[62521]: DEBUG nova.compute.manager [req-7bec6aee-d58c-4597-a802-50d794de292d req-f5bae662-b21f-4a43-8502-7c917c26c897 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] No waiting events found dispatching network-vif-plugged-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.197501] env[62521]: WARNING nova.compute.manager [req-7bec6aee-d58c-4597-a802-50d794de292d req-f5bae662-b21f-4a43-8502-7c917c26c897 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received unexpected event network-vif-plugged-72671c1a-eabd-4a80-a591-17b26cea0249 for instance with vm_state building and task_state spawning. [ 846.231755] env[62521]: DEBUG nova.network.neutron [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Successfully updated port: 72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 846.257296] env[62521]: DEBUG oslo_concurrency.lockutils [req-c153a47c-03db-4794-9194-650c03b0932f req-43a25569-0426-4550-8f2a-065cbe3fc13c service nova] Releasing lock "refresh_cache-583b1a1b-1817-45fc-aa4f-fc9de4b4a243" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.353716] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526fcf0a-0b67-8e5d-3d20-2532432a85b2, 'name': SearchDatastore_Task, 'duration_secs': 0.011923} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.357021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.357021] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.357021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.357021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.357021] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.357021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.357021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.357021] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-068db3d4-e11e-4497-935a-8a18357751f8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.360791] env[62521]: DEBUG oslo_concurrency.lockutils [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquiring lock "4afe09db-6c01-444f-a127-6e1f97794544" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.361139] env[62521]: DEBUG oslo_concurrency.lockutils [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lock "4afe09db-6c01-444f-a127-6e1f97794544" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.361494] env[62521]: DEBUG oslo_concurrency.lockutils [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquiring lock "4afe09db-6c01-444f-a127-6e1f97794544-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.361815] env[62521]: DEBUG oslo_concurrency.lockutils [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lock "4afe09db-6c01-444f-a127-6e1f97794544-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.362131] env[62521]: DEBUG oslo_concurrency.lockutils [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lock "4afe09db-6c01-444f-a127-6e1f97794544-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.363712] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fed31fd4-b688-4eee-9664-f1fe36b892b6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.365975] env[62521]: INFO nova.compute.manager [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Terminating instance [ 846.368391] env[62521]: DEBUG nova.compute.manager [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.368783] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 846.370020] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61abd55b-0bc0-410a-915f-37cd681f0642 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.374156] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 846.374156] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ab685a-78da-f288-5d06-cf05f7adda10" [ 846.374156] env[62521]: _type = "Task" [ 846.374156] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.380624] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.381086] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.381369] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.382462] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a2a96ec-9785-4cd8-845c-52ce169897dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.383961] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79f025c9-5bf1-4254-a7ae-97a2382be407 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.388986] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ab685a-78da-f288-5d06-cf05f7adda10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.392207] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 846.392207] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522d63f3-9ba2-6fb8-fc30-d65ab8b63378" [ 846.392207] env[62521]: _type = "Task" [ 846.392207] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.396842] env[62521]: DEBUG oslo_vmware.api [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for the task: (returnval){ [ 846.396842] env[62521]: value = "task-1318555" [ 846.396842] env[62521]: _type = "Task" [ 846.396842] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.407025] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522d63f3-9ba2-6fb8-fc30-d65ab8b63378, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.414272] env[62521]: DEBUG nova.compute.utils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.424986] env[62521]: DEBUG oslo_vmware.api [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318555, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.424986] env[62521]: DEBUG nova.compute.manager [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.424986] env[62521]: DEBUG nova.network.neutron [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.443021] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318554, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083807} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.443021] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.443021] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9329b2e-d5ed-4d1b-bc02-d827acd0ad8c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.464598] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 92629ebf-eb50-4862-abe2-01e210e302ad/92629ebf-eb50-4862-abe2-01e210e302ad.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.466071] env[62521]: DEBUG nova.policy [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '951010665d8e4908b7d527ec0f97fa52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ccd393a0715144e8af991f0b304bb142', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.468340] env[62521]: DEBUG nova.network.neutron [req-a7ffb8de-ea30-4a92-9286-2765eb5cf4be req-844c2bff-6441-46ec-b159-209fd8fe3a8b service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updated VIF entry in instance network info cache for port 903b74d0-11d2-404d-81f1-b6bb18f713af. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 846.468679] env[62521]: DEBUG nova.network.neutron [req-a7ffb8de-ea30-4a92-9286-2765eb5cf4be req-844c2bff-6441-46ec-b159-209fd8fe3a8b service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updating instance_info_cache with network_info: [{"id": "903b74d0-11d2-404d-81f1-b6bb18f713af", "address": "fa:16:3e:6a:82:e1", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap903b74d0-11", "ovs_interfaceid": "903b74d0-11d2-404d-81f1-b6bb18f713af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.473019] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e857fd8-d01f-4e26-a6ce-e6029582d656 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.487923] env[62521]: DEBUG oslo_concurrency.lockutils [req-a7ffb8de-ea30-4a92-9286-2765eb5cf4be req-844c2bff-6441-46ec-b159-209fd8fe3a8b service nova] Releasing lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.494467] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 846.494467] env[62521]: value = "task-1318556" [ 846.494467] env[62521]: _type = "Task" [ 846.494467] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.503863] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318556, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.566121] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cc5904-643f-e61a-6e06-b801994aeb66, 'name': SearchDatastore_Task, 'duration_secs': 0.032946} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.566462] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.566678] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.566912] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.567085] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.567276] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.567541] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b6990da-1c7f-41d3-a196-9caf2a6dfc09 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.576919] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.577157] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.577891] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02714947-1822-485d-b8cc-5aa5fca7513d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.583118] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 846.583118] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52409402-196a-5d45-136a-2ee02cb90a2c" [ 846.583118] env[62521]: _type = "Task" [ 846.583118] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.590971] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52409402-196a-5d45-136a-2ee02cb90a2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.601122] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318553, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.735671] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.735671] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.735671] env[62521]: DEBUG nova.network.neutron [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 846.885966] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ab685a-78da-f288-5d06-cf05f7adda10, 'name': SearchDatastore_Task, 'duration_secs': 0.026198} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.886337] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.886569] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.886780] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.902676] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522d63f3-9ba2-6fb8-fc30-d65ab8b63378, 'name': SearchDatastore_Task, 'duration_secs': 0.020698} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.907621] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4800681-d0e2-41ed-b41f-16072dd513d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.915980] env[62521]: DEBUG oslo_vmware.api [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318555, 'name': PowerOffVM_Task, 'duration_secs': 0.418809} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.917688] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 846.917688] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 846.917885] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 846.917885] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d56772-7eb7-b29c-f129-7a12136a8048" [ 846.917885] env[62521]: _type = "Task" [ 846.917885] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.918143] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fd2518e-6a3f-44c9-bff7-1de8ad197bf4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.929061] env[62521]: DEBUG nova.compute.manager [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.939513] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d56772-7eb7-b29c-f129-7a12136a8048, 'name': SearchDatastore_Task, 'duration_secs': 0.011524} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.939699] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.939915] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6/70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.940290] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.940495] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.940774] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e668081-0bed-43bb-9722-33df44745b86 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.943966] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87af4c9f-26d9-4e05-9ba8-6b0ec82598f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.954341] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 846.954341] env[62521]: value = "task-1318558" [ 846.954341] env[62521]: _type = "Task" [ 846.954341] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.963139] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.963651] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.965438] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3ce6cfc-d2b1-4956-8921-c6902ac8b1ab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.978537] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.993020] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 846.993020] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 846.993020] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Deleting the datastore file [datastore2] 4afe09db-6c01-444f-a127-6e1f97794544 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.993020] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7de46712-055d-4428-9903-59998fa78331 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.993020] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 846.993020] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5298e1c0-11e5-f3f8-f4e8-c8a44088f5cb" [ 846.993020] env[62521]: _type = "Task" [ 846.993020] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.000913] env[62521]: DEBUG oslo_vmware.api [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for the task: (returnval){ [ 847.000913] env[62521]: value = "task-1318559" [ 847.000913] env[62521]: _type = "Task" [ 847.000913] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.029297] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5298e1c0-11e5-f3f8-f4e8-c8a44088f5cb, 'name': SearchDatastore_Task, 'duration_secs': 0.015083} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.034427] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ce609c3-ce87-4969-8acc-299e1779f31a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.038433] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318556, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.047394] env[62521]: DEBUG oslo_vmware.api [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318559, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.051560] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 847.051560] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5225dc02-19ed-d8d5-46c3-4e08553391f2" [ 847.051560] env[62521]: _type = "Task" [ 847.051560] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.058337] env[62521]: DEBUG nova.network.neutron [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Successfully created port: 20597486-a59c-4869-9c6e-3d10e7b83d5f {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.070854] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5225dc02-19ed-d8d5-46c3-4e08553391f2, 'name': SearchDatastore_Task, 'duration_secs': 0.009912} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.071180] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.071471] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 583b1a1b-1817-45fc-aa4f-fc9de4b4a243/583b1a1b-1817-45fc-aa4f-fc9de4b4a243.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.071845] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d9e6282-9c5b-4571-9a12-edd9ef57c2fe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.079677] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 847.079677] env[62521]: value = "task-1318560" [ 847.079677] env[62521]: _type = "Task" [ 847.079677] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.094848] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318560, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.101832] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52409402-196a-5d45-136a-2ee02cb90a2c, 'name': SearchDatastore_Task, 'duration_secs': 0.019679} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.106144] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5e25eac-d9c3-4c88-8193-84ff7a7275fb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.112738] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318553, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.119105] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 847.119105] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e67e2e-70a9-3bc8-07de-6f0c6d7c9a06" [ 847.119105] env[62521]: _type = "Task" [ 847.119105] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.127148] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e67e2e-70a9-3bc8-07de-6f0c6d7c9a06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.271469] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1bcecf-f55d-428a-8c04-c91cf6851eb6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.280123] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1796a53-6cb8-4d06-b7ba-6db6d5026e0f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.284770] env[62521]: DEBUG nova.network.neutron [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.321756] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909ddb7a-095b-4b4d-8de7-b65bcd0f6755 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.331000] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab4e441-c8fb-490e-8ca4-3d0a503b4a2d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.348171] env[62521]: DEBUG nova.compute.provider_tree [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.467286] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318558, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.473750] env[62521]: DEBUG nova.network.neutron [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating instance_info_cache with network_info: [{"id": "72671c1a-eabd-4a80-a591-17b26cea0249", "address": "fa:16:3e:3a:1f:96", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72671c1a-ea", "ovs_interfaceid": "72671c1a-eabd-4a80-a591-17b26cea0249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.510179] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318556, 'name': ReconfigVM_Task, 'duration_secs': 0.542584} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.513560] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 92629ebf-eb50-4862-abe2-01e210e302ad/92629ebf-eb50-4862-abe2-01e210e302ad.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.514250] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb06cdcc-50f9-4070-8dc6-0c6f4e5ca6fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.524193] env[62521]: DEBUG oslo_vmware.api [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Task: {'id': task-1318559, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.380661} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.525281] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 847.525481] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 847.525663] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 847.525893] env[62521]: INFO nova.compute.manager [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Took 1.16 seconds to destroy the instance on the hypervisor. [ 847.526158] env[62521]: DEBUG oslo.service.loopingcall [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.526439] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 847.526439] env[62521]: value = "task-1318561" [ 847.526439] env[62521]: _type = "Task" [ 847.526439] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.526907] env[62521]: DEBUG nova.compute.manager [-] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.527011] env[62521]: DEBUG nova.network.neutron [-] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 847.541786] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318561, 'name': Rename_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.595398] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318560, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.608189] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318553, 'name': CloneVM_Task, 'duration_secs': 1.751972} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.609047] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Created linked-clone VM from snapshot [ 847.609330] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b48bede-ec16-4f6c-a6c8-09b444f00a75 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.617709] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Uploading image 9f64b520-b041-469a-8080-6b597c951678 {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 847.629971] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e67e2e-70a9-3bc8-07de-6f0c6d7c9a06, 'name': SearchDatastore_Task, 'duration_secs': 0.026786} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.630290] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.630553] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0/bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.630830] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c7401b9e-70c6-4f9a-81d9-aeadc972ea1e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.642106] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 847.642106] env[62521]: value = "task-1318562" [ 847.642106] env[62521]: _type = "Task" [ 847.642106] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.648841] env[62521]: DEBUG oslo_vmware.rw_handles [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 847.648841] env[62521]: value = "vm-282094" [ 847.648841] env[62521]: _type = "VirtualMachine" [ 847.648841] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 847.648841] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-da5c1df1-fd4f-4007-a1e2-f61cf6643be3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.653977] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318562, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.660424] env[62521]: DEBUG oslo_vmware.rw_handles [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lease: (returnval){ [ 847.660424] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529a49a8-15f5-1ef2-958b-81d0b06d8580" [ 847.660424] env[62521]: _type = "HttpNfcLease" [ 847.660424] env[62521]: } obtained for exporting VM: (result){ [ 847.660424] env[62521]: value = "vm-282094" [ 847.660424] env[62521]: _type = "VirtualMachine" [ 847.660424] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 847.660732] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the lease: (returnval){ [ 847.660732] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529a49a8-15f5-1ef2-958b-81d0b06d8580" [ 847.660732] env[62521]: _type = "HttpNfcLease" [ 847.660732] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 847.670043] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 847.670043] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529a49a8-15f5-1ef2-958b-81d0b06d8580" [ 847.670043] env[62521]: _type = "HttpNfcLease" [ 847.670043] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 847.851884] env[62521]: DEBUG nova.scheduler.client.report [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.884354] env[62521]: DEBUG nova.compute.manager [req-f8191f29-4f6a-4960-881e-f180d3e49cd9 req-d16d6c4c-9eb9-42df-b18d-6cf996fcc238 service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Received event network-vif-deleted-b520981a-8f32-4f58-9e84-a062b20d933a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.884569] env[62521]: INFO nova.compute.manager [req-f8191f29-4f6a-4960-881e-f180d3e49cd9 req-d16d6c4c-9eb9-42df-b18d-6cf996fcc238 service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Neutron deleted interface b520981a-8f32-4f58-9e84-a062b20d933a; detaching it from the instance and deleting it from the info cache [ 847.884768] env[62521]: DEBUG nova.network.neutron [req-f8191f29-4f6a-4960-881e-f180d3e49cd9 req-d16d6c4c-9eb9-42df-b18d-6cf996fcc238 service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.936945] env[62521]: DEBUG nova.compute.manager [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.958379] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.958731] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.958799] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.958957] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.959816] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.959816] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.959816] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.959816] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.959816] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.960045] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.960078] env[62521]: DEBUG nova.virt.hardware [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.960960] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903fc6db-9e89-4257-bfd1-cf4edfac62f8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.975893] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793e61c2-3b6d-442b-850f-02f73d5a4f29 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.980831] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534558} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.981328] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.981629] env[62521]: DEBUG nova.compute.manager [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Instance network_info: |[{"id": "72671c1a-eabd-4a80-a591-17b26cea0249", "address": "fa:16:3e:3a:1f:96", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72671c1a-ea", "ovs_interfaceid": "72671c1a-eabd-4a80-a591-17b26cea0249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 847.981894] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6/70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 847.982113] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.982482] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:1f:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ec3f9e71-839a-429d-b211-d3dfc98ca4f6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72671c1a-eabd-4a80-a591-17b26cea0249', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 847.992471] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating folder: Project (6ab001785c25414abbaa4bfe3a6a0bb4). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 847.993164] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bbe70d91-ed0e-4650-9440-3275d07a88b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.003971] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-154c4e64-8138-4fba-a31d-c722e6213025 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.010344] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 848.010344] env[62521]: value = "task-1318565" [ 848.010344] env[62521]: _type = "Task" [ 848.010344] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.015042] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Created folder: Project (6ab001785c25414abbaa4bfe3a6a0bb4) in parent group-v282025. [ 848.015676] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating folder: Instances. Parent ref: group-v282095. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.015815] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c228491f-5c57-4d98-bf86-76ad35e39b29 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.021273] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318565, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.029498] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Created folder: Instances in parent group-v282095. [ 848.030467] env[62521]: DEBUG oslo.service.loopingcall [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.030467] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 848.030603] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d75d0b02-e6df-4604-a014-671adefc443f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.055370] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318561, 'name': Rename_Task, 'duration_secs': 0.278143} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.056762] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.057015] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.057015] env[62521]: value = "task-1318567" [ 848.057015] env[62521]: _type = "Task" [ 848.057015] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.057224] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8b976b3-2c7f-426a-ba2d-d5645106dd72 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.066925] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318567, 'name': CreateVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.068265] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 848.068265] env[62521]: value = "task-1318568" [ 848.068265] env[62521]: _type = "Task" [ 848.068265] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.075817] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318568, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.091456] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318560, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.831398} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.091699] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 583b1a1b-1817-45fc-aa4f-fc9de4b4a243/583b1a1b-1817-45fc-aa4f-fc9de4b4a243.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 848.091906] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.092161] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57ca7c0e-8d07-4eb5-a1ac-12e559de7e45 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.098205] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 848.098205] env[62521]: value = "task-1318569" [ 848.098205] env[62521]: _type = "Task" [ 848.098205] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.106651] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318569, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.154217] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318562, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.169528] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 848.169528] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529a49a8-15f5-1ef2-958b-81d0b06d8580" [ 848.169528] env[62521]: _type = "HttpNfcLease" [ 848.169528] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 848.169831] env[62521]: DEBUG oslo_vmware.rw_handles [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 848.169831] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529a49a8-15f5-1ef2-958b-81d0b06d8580" [ 848.169831] env[62521]: _type = "HttpNfcLease" [ 848.169831] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 848.170713] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abcadd4-e595-4fba-b167-2c5f7cee21c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.180765] env[62521]: DEBUG oslo_vmware.rw_handles [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec15df-88a4-bf8a-634b-31965c48977a/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 848.180994] env[62521]: DEBUG oslo_vmware.rw_handles [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec15df-88a4-bf8a-634b-31965c48977a/disk-0.vmdk for reading. {{(pid=62521) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 848.300865] env[62521]: DEBUG nova.compute.manager [req-b3025d73-582d-4abd-8655-3f1ebb4edfe4 req-c45b9208-1d47-46d6-a7d9-f2c541cb8f63 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received event network-changed-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.301233] env[62521]: DEBUG nova.compute.manager [req-b3025d73-582d-4abd-8655-3f1ebb4edfe4 req-c45b9208-1d47-46d6-a7d9-f2c541cb8f63 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Refreshing instance network info cache due to event network-changed-72671c1a-eabd-4a80-a591-17b26cea0249. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 848.301563] env[62521]: DEBUG oslo_concurrency.lockutils [req-b3025d73-582d-4abd-8655-3f1ebb4edfe4 req-c45b9208-1d47-46d6-a7d9-f2c541cb8f63 service nova] Acquiring lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.301796] env[62521]: DEBUG oslo_concurrency.lockutils [req-b3025d73-582d-4abd-8655-3f1ebb4edfe4 req-c45b9208-1d47-46d6-a7d9-f2c541cb8f63 service nova] Acquired lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.302133] env[62521]: DEBUG nova.network.neutron [req-b3025d73-582d-4abd-8655-3f1ebb4edfe4 req-c45b9208-1d47-46d6-a7d9-f2c541cb8f63 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Refreshing network info cache for port 72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 848.325644] env[62521]: DEBUG nova.network.neutron [-] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.361709] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.362040] env[62521]: DEBUG nova.compute.manager [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.366978] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a7d5c0af-a68b-47fd-afdf-4b67c89b6f79 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.369144] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.475s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.369479] env[62521]: DEBUG nova.objects.instance [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lazy-loading 'resources' on Instance uuid 062f311c-a783-41cf-8d83-b3a8d4df14f4 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.388145] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb134674-b7d8-4006-9fe1-9f143e656f22 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.403194] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ca98e9-7744-492b-b65a-fa1c65edd66a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.446671] env[62521]: DEBUG nova.compute.manager [req-f8191f29-4f6a-4960-881e-f180d3e49cd9 req-d16d6c4c-9eb9-42df-b18d-6cf996fcc238 service nova] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Detach interface failed, port_id=b520981a-8f32-4f58-9e84-a062b20d933a, reason: Instance 4afe09db-6c01-444f-a127-6e1f97794544 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 848.522497] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318565, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064837} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.522906] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.524116] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b2e34d-8ade-454f-bc3b-1003c84d9343 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.554793] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6/70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.555172] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ca75d03-6dff-4bce-a633-6ff1277ac759 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.594018] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318567, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.596256] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 848.596256] env[62521]: value = "task-1318570" [ 848.596256] env[62521]: _type = "Task" [ 848.596256] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.596471] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318568, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.610563] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.617225] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318569, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119099} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.619924] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.619924] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc05846-0781-41e1-831d-56745a028a22 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.655523] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 583b1a1b-1817-45fc-aa4f-fc9de4b4a243/583b1a1b-1817-45fc-aa4f-fc9de4b4a243.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.664881] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7778fe5-4dbe-4987-9c4f-c70680f811e5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.687378] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318562, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.949982} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.689195] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0/bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 848.689538] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.689996] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 848.689996] env[62521]: value = "task-1318571" [ 848.689996] env[62521]: _type = "Task" [ 848.689996] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.690282] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6aeb7ff-c6c5-4acc-914b-28ce36411440 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.702309] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318571, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.704213] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 848.704213] env[62521]: value = "task-1318572" [ 848.704213] env[62521]: _type = "Task" [ 848.704213] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.714408] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318572, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.829766] env[62521]: INFO nova.compute.manager [-] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Took 1.30 seconds to deallocate network for instance. [ 848.861139] env[62521]: DEBUG nova.network.neutron [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Successfully updated port: 20597486-a59c-4869-9c6e-3d10e7b83d5f {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.874022] env[62521]: DEBUG nova.compute.utils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.877097] env[62521]: DEBUG nova.compute.manager [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.877466] env[62521]: DEBUG nova.network.neutron [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.966789] env[62521]: DEBUG nova.policy [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a0e35b0e4574ed9afc4549d30daa210', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '31c788848b0d47478564e53066e7c51a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 849.071306] env[62521]: DEBUG nova.network.neutron [req-b3025d73-582d-4abd-8655-3f1ebb4edfe4 req-c45b9208-1d47-46d6-a7d9-f2c541cb8f63 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updated VIF entry in instance network info cache for port 72671c1a-eabd-4a80-a591-17b26cea0249. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 849.072763] env[62521]: DEBUG nova.network.neutron [req-b3025d73-582d-4abd-8655-3f1ebb4edfe4 req-c45b9208-1d47-46d6-a7d9-f2c541cb8f63 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating instance_info_cache with network_info: [{"id": "72671c1a-eabd-4a80-a591-17b26cea0249", "address": "fa:16:3e:3a:1f:96", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72671c1a-ea", "ovs_interfaceid": "72671c1a-eabd-4a80-a591-17b26cea0249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.093052] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318567, 'name': CreateVM_Task, 'duration_secs': 0.615279} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.099024] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.099024] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318568, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.099024] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.099024] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.099024] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.099024] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-825d499b-de43-4072-830d-e0d8d3d61b6d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.110401] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 849.110401] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529c4232-5b23-5c04-a7c5-b20bf5bb5297" [ 849.110401] env[62521]: _type = "Task" [ 849.110401] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.114349] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.123933] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529c4232-5b23-5c04-a7c5-b20bf5bb5297, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.171846] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e412f7f-bd0a-428f-a4a2-f3f5a588dafb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.182718] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90edbdb-3d85-43be-b4a2-8f61dbfa356e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.231581] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e19f0e-fa25-47a4-ae34-1904a2695ad1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.243107] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318571, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.249334] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318572, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109055} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.250749] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdacd02-5869-4b6a-8226-6ee89c18dd12 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.256370] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.257322] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40fc6724-ede0-4fec-b57e-3800b8cbf4e2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.285150] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0/bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.293321] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc671d28-3cb0-4a2a-9df1-a5c211c49474 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.309309] env[62521]: DEBUG nova.compute.provider_tree [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.318558] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 849.318558] env[62521]: value = "task-1318573" [ 849.318558] env[62521]: _type = "Task" [ 849.318558] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.326412] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318573, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.338563] env[62521]: DEBUG oslo_concurrency.lockutils [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.362256] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquiring lock "refresh_cache-2a0ccda4-87a6-4bb3-bf34-4434faf574e6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.362256] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquired lock "refresh_cache-2a0ccda4-87a6-4bb3-bf34-4434faf574e6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.362256] env[62521]: DEBUG nova.network.neutron [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.381659] env[62521]: DEBUG nova.compute.manager [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.577428] env[62521]: DEBUG oslo_concurrency.lockutils [req-b3025d73-582d-4abd-8655-3f1ebb4edfe4 req-c45b9208-1d47-46d6-a7d9-f2c541cb8f63 service nova] Releasing lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.592435] env[62521]: DEBUG oslo_vmware.api [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318568, 'name': PowerOnVM_Task, 'duration_secs': 1.367458} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.592761] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.592993] env[62521]: INFO nova.compute.manager [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Took 8.81 seconds to spawn the instance on the hypervisor. [ 849.593433] env[62521]: DEBUG nova.compute.manager [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.594411] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83140531-0bdc-461a-9768-3233280c121e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.615832] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318570, 'name': ReconfigVM_Task, 'duration_secs': 0.837565} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.615832] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6/70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.618175] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-506886cc-82e3-4d2c-a95f-5c45db598b87 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.629377] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 849.629377] env[62521]: value = "task-1318574" [ 849.629377] env[62521]: _type = "Task" [ 849.629377] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.629634] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529c4232-5b23-5c04-a7c5-b20bf5bb5297, 'name': SearchDatastore_Task, 'duration_secs': 0.017152} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.630020] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.630247] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 849.631154] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.631236] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.632103] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 849.635021] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d89c5eff-ac62-4e58-8956-72ffefb6b2bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.637601] env[62521]: DEBUG nova.network.neutron [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Successfully created port: 8824a231-01df-41e5-9bea-9071d8b41cb8 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.645489] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318574, 'name': Rename_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.647298] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 849.647703] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 849.648507] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a478c7a3-e7d5-4a75-ab13-844374aea039 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.654300] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 849.654300] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f3ae68-c2f5-eb55-757b-24c11979adc9" [ 849.654300] env[62521]: _type = "Task" [ 849.654300] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.662832] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f3ae68-c2f5-eb55-757b-24c11979adc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.734128] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318571, 'name': ReconfigVM_Task, 'duration_secs': 0.756703} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.734518] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 583b1a1b-1817-45fc-aa4f-fc9de4b4a243/583b1a1b-1817-45fc-aa4f-fc9de4b4a243.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.735239] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3f40a649-ccae-44d0-817b-222082f555ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.742739] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 849.742739] env[62521]: value = "task-1318575" [ 849.742739] env[62521]: _type = "Task" [ 849.742739] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.752161] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318575, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.813124] env[62521]: DEBUG nova.scheduler.client.report [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.828313] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.912680] env[62521]: DEBUG nova.network.neutron [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.084396] env[62521]: DEBUG nova.network.neutron [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Updating instance_info_cache with network_info: [{"id": "20597486-a59c-4869-9c6e-3d10e7b83d5f", "address": "fa:16:3e:00:cb:7f", "network": {"id": "82b102b4-acc7-48c9-a665-7771e28b75c2", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1256938968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccd393a0715144e8af991f0b304bb142", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20597486-a5", "ovs_interfaceid": "20597486-a59c-4869-9c6e-3d10e7b83d5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.123321] env[62521]: INFO nova.compute.manager [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Took 28.55 seconds to build instance. [ 850.140228] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318574, 'name': Rename_Task, 'duration_secs': 0.223832} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.140551] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.140815] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f982660c-5f9e-4efe-8189-bdd3c2b022ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.148344] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 850.148344] env[62521]: value = "task-1318576" [ 850.148344] env[62521]: _type = "Task" [ 850.148344] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.157473] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318576, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.166814] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f3ae68-c2f5-eb55-757b-24c11979adc9, 'name': SearchDatastore_Task, 'duration_secs': 0.012753} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.167759] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a518190-8cb0-410a-b5cd-4246925fc680 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.172925] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 850.172925] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bc36d8-a7b5-838d-72cc-8a78afde17f4" [ 850.172925] env[62521]: _type = "Task" [ 850.172925] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.184411] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bc36d8-a7b5-838d-72cc-8a78afde17f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.254782] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318575, 'name': Rename_Task, 'duration_secs': 0.204382} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.254782] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.254782] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c35cda1-099e-4a9b-b040-028be89912c3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.265028] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 850.265028] env[62521]: value = "task-1318577" [ 850.265028] env[62521]: _type = "Task" [ 850.265028] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.270267] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.323561] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.954s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.326541] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 14.300s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.334902] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318573, 'name': ReconfigVM_Task, 'duration_secs': 0.612604} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.335464] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Reconfigured VM instance instance-00000041 to attach disk [datastore1] bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0/bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.336413] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-482ee1a2-eec1-49fa-a0d2-375c85a95263 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.346021] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 850.346021] env[62521]: value = "task-1318578" [ 850.346021] env[62521]: _type = "Task" [ 850.346021] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.350791] env[62521]: INFO nova.scheduler.client.report [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Deleted allocations for instance 062f311c-a783-41cf-8d83-b3a8d4df14f4 [ 850.358608] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318578, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.395088] env[62521]: DEBUG nova.compute.manager [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.426346] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.426645] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.426774] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.426965] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.427136] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.427291] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.427504] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.427667] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.427840] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.428075] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.428247] env[62521]: DEBUG nova.virt.hardware [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.429458] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45395767-fab9-4580-8e6d-7960f3e00747 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.439114] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757bb86d-13f6-4b89-b4cc-0d64013b0512 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.465028] env[62521]: DEBUG nova.compute.manager [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Received event network-vif-plugged-20597486-a59c-4869-9c6e-3d10e7b83d5f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.465384] env[62521]: DEBUG oslo_concurrency.lockutils [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] Acquiring lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.465474] env[62521]: DEBUG oslo_concurrency.lockutils [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] Lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.465633] env[62521]: DEBUG oslo_concurrency.lockutils [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] Lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.465852] env[62521]: DEBUG nova.compute.manager [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] No waiting events found dispatching network-vif-plugged-20597486-a59c-4869-9c6e-3d10e7b83d5f {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.465942] env[62521]: WARNING nova.compute.manager [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Received unexpected event network-vif-plugged-20597486-a59c-4869-9c6e-3d10e7b83d5f for instance with vm_state building and task_state spawning. [ 850.466129] env[62521]: DEBUG nova.compute.manager [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Received event network-changed-20597486-a59c-4869-9c6e-3d10e7b83d5f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.466296] env[62521]: DEBUG nova.compute.manager [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Refreshing instance network info cache due to event network-changed-20597486-a59c-4869-9c6e-3d10e7b83d5f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.466467] env[62521]: DEBUG oslo_concurrency.lockutils [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] Acquiring lock "refresh_cache-2a0ccda4-87a6-4bb3-bf34-4434faf574e6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.587470] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Releasing lock "refresh_cache-2a0ccda4-87a6-4bb3-bf34-4434faf574e6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.587905] env[62521]: DEBUG nova.compute.manager [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Instance network_info: |[{"id": "20597486-a59c-4869-9c6e-3d10e7b83d5f", "address": "fa:16:3e:00:cb:7f", "network": {"id": "82b102b4-acc7-48c9-a665-7771e28b75c2", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1256938968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccd393a0715144e8af991f0b304bb142", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20597486-a5", "ovs_interfaceid": "20597486-a59c-4869-9c6e-3d10e7b83d5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.588355] env[62521]: DEBUG oslo_concurrency.lockutils [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] Acquired lock "refresh_cache-2a0ccda4-87a6-4bb3-bf34-4434faf574e6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.588601] env[62521]: DEBUG nova.network.neutron [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Refreshing network info cache for port 20597486-a59c-4869-9c6e-3d10e7b83d5f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.590263] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:cb:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0248a27a-1d7f-4195-987b-06bfc8467347', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '20597486-a59c-4869-9c6e-3d10e7b83d5f', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.600931] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Creating folder: Project (ccd393a0715144e8af991f0b304bb142). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.605358] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8455361-2478-42c6-a82a-6b8586a70a21 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.618026] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Created folder: Project (ccd393a0715144e8af991f0b304bb142) in parent group-v282025. [ 850.618026] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Creating folder: Instances. Parent ref: group-v282098. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.618476] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f52ba9ce-d0d2-4427-96e1-0fd9742cf764 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.631154] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1abc21fc-1971-4d90-8625-171c70c8b567 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "92629ebf-eb50-4862-abe2-01e210e302ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.240s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.631154] env[62521]: DEBUG oslo_concurrency.lockutils [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquiring lock "92629ebf-eb50-4862-abe2-01e210e302ad" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.631154] env[62521]: DEBUG oslo_concurrency.lockutils [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "92629ebf-eb50-4862-abe2-01e210e302ad" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.631933] env[62521]: INFO nova.compute.manager [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Rebooting instance [ 850.634422] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Created folder: Instances in parent group-v282098. [ 850.636124] env[62521]: DEBUG oslo.service.loopingcall [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.636124] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.636485] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31a77465-7c9a-468c-9761-134b52cd031d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.673725] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318576, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.678991] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.678991] env[62521]: value = "task-1318581" [ 850.678991] env[62521]: _type = "Task" [ 850.678991] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.688835] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bc36d8-a7b5-838d-72cc-8a78afde17f4, 'name': SearchDatastore_Task, 'duration_secs': 0.020445} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.688835] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.689042] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad/73bc1cf1-7ee1-4a9b-8214-2dc752b700ad.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 850.689715] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-537b67a0-2f79-48fa-b342-66c11e3f64d0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.694345] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318581, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.699765] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 850.699765] env[62521]: value = "task-1318582" [ 850.699765] env[62521]: _type = "Task" [ 850.699765] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.708482] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318582, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.773087] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318577, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.834695] env[62521]: INFO nova.compute.claims [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.855418] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318578, 'name': Rename_Task, 'duration_secs': 0.233825} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.855616] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.855900] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a95e205-874b-48a3-aef9-bcb593bdea88 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.864554] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 850.864554] env[62521]: value = "task-1318583" [ 850.864554] env[62521]: _type = "Task" [ 850.864554] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.865089] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d24c0cae-5c21-40fa-95b8-a3cc1987e377 tempest-ServerShowV247Test-353971570 tempest-ServerShowV247Test-353971570-project-member] Lock "062f311c-a783-41cf-8d83-b3a8d4df14f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.757s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.876333] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.883860] env[62521]: DEBUG nova.network.neutron [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Updated VIF entry in instance network info cache for port 20597486-a59c-4869-9c6e-3d10e7b83d5f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 850.884280] env[62521]: DEBUG nova.network.neutron [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Updating instance_info_cache with network_info: [{"id": "20597486-a59c-4869-9c6e-3d10e7b83d5f", "address": "fa:16:3e:00:cb:7f", "network": {"id": "82b102b4-acc7-48c9-a665-7771e28b75c2", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1256938968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccd393a0715144e8af991f0b304bb142", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20597486-a5", "ovs_interfaceid": "20597486-a59c-4869-9c6e-3d10e7b83d5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.162123] env[62521]: DEBUG oslo_concurrency.lockutils [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquiring lock "refresh_cache-92629ebf-eb50-4862-abe2-01e210e302ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.162424] env[62521]: DEBUG oslo_concurrency.lockutils [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquired lock "refresh_cache-92629ebf-eb50-4862-abe2-01e210e302ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.162626] env[62521]: DEBUG nova.network.neutron [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.174036] env[62521]: DEBUG oslo_vmware.api [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318576, 'name': PowerOnVM_Task, 'duration_secs': 0.552072} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.175113] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.175364] env[62521]: DEBUG nova.compute.manager [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.176332] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdcd72b-8eb8-4bb0-bb19-4548b4a2f561 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.193366] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318581, 'name': CreateVM_Task, 'duration_secs': 0.42912} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.193845] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.194659] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.194831] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.195232] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.195517] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e46b503-9fcf-45e1-952d-49005334252b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.202743] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for the task: (returnval){ [ 851.202743] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ae68d0-0183-69f6-2b79-b27f57c75e21" [ 851.202743] env[62521]: _type = "Task" [ 851.202743] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.215745] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318582, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.220037] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ae68d0-0183-69f6-2b79-b27f57c75e21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.277797] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318577, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.343565] env[62521]: INFO nova.compute.resource_tracker [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating resource usage from migration 29ec3101-66e2-4662-a2ab-f405aeb20552 [ 851.377890] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318583, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.386595] env[62521]: DEBUG oslo_concurrency.lockutils [req-90bd28bc-d13e-4426-8d42-9704dd6b26a6 req-f7bd9950-1364-4d32-a145-065036545d60 service nova] Releasing lock "refresh_cache-2a0ccda4-87a6-4bb3-bf34-4434faf574e6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.566034] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736e73d6-42df-4f76-85a9-8a1b774aacbe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.574744] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3190f505-183c-4f2f-b274-5f2a12e2bd81 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.612050] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f3219f-d42f-42d4-9bc0-de5342f1e966 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.622180] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab4a6c8-048e-465a-ab5d-ce1dcbce674c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.642828] env[62521]: DEBUG nova.compute.provider_tree [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.708841] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.725221] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318582, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597158} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.726589] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad/73bc1cf1-7ee1-4a9b-8214-2dc752b700ad.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 851.726855] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.728224] env[62521]: DEBUG nova.compute.manager [req-7a1b7f0d-f6ca-4b31-aaca-bc94f49d6606 req-a14d1f46-4167-4d91-91fe-3c3537897211 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Received event network-vif-plugged-8824a231-01df-41e5-9bea-9071d8b41cb8 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.730322] env[62521]: DEBUG oslo_concurrency.lockutils [req-7a1b7f0d-f6ca-4b31-aaca-bc94f49d6606 req-a14d1f46-4167-4d91-91fe-3c3537897211 service nova] Acquiring lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.730322] env[62521]: DEBUG oslo_concurrency.lockutils [req-7a1b7f0d-f6ca-4b31-aaca-bc94f49d6606 req-a14d1f46-4167-4d91-91fe-3c3537897211 service nova] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.730322] env[62521]: DEBUG oslo_concurrency.lockutils [req-7a1b7f0d-f6ca-4b31-aaca-bc94f49d6606 req-a14d1f46-4167-4d91-91fe-3c3537897211 service nova] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.730322] env[62521]: DEBUG nova.compute.manager [req-7a1b7f0d-f6ca-4b31-aaca-bc94f49d6606 req-a14d1f46-4167-4d91-91fe-3c3537897211 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] No waiting events found dispatching network-vif-plugged-8824a231-01df-41e5-9bea-9071d8b41cb8 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 851.730322] env[62521]: WARNING nova.compute.manager [req-7a1b7f0d-f6ca-4b31-aaca-bc94f49d6606 req-a14d1f46-4167-4d91-91fe-3c3537897211 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Received unexpected event network-vif-plugged-8824a231-01df-41e5-9bea-9071d8b41cb8 for instance with vm_state building and task_state spawning. [ 851.736395] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39b01825-76b1-4272-9674-83b69ceaf56e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.739756] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ae68d0-0183-69f6-2b79-b27f57c75e21, 'name': SearchDatastore_Task, 'duration_secs': 0.027438} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.740182] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.740513] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.740780] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.740981] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.741197] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.741991] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b152ecbb-1fa8-4e91-ac70-e0dac66543a3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.747499] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 851.747499] env[62521]: value = "task-1318584" [ 851.747499] env[62521]: _type = "Task" [ 851.747499] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.753743] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.753953] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.757721] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83839aa9-2291-4ca6-8f91-f09bb6e0c03d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.760563] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318584, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.764823] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for the task: (returnval){ [ 851.764823] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529a40fc-bc8d-3f49-dad1-54db3f0cda4d" [ 851.764823] env[62521]: _type = "Task" [ 851.764823] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.776706] env[62521]: DEBUG oslo_vmware.api [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318577, 'name': PowerOnVM_Task, 'duration_secs': 1.485833} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.782183] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.782879] env[62521]: INFO nova.compute.manager [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Took 15.17 seconds to spawn the instance on the hypervisor. [ 851.782879] env[62521]: DEBUG nova.compute.manager [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.783267] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529a40fc-bc8d-3f49-dad1-54db3f0cda4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.784066] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29bf5c9-c75d-4327-8963-771b7be87e25 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.879994] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318583, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.973043] env[62521]: DEBUG nova.network.neutron [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Updating instance_info_cache with network_info: [{"id": "892c26c7-9b46-4a0f-a406-6bdc9c0872fa", "address": "fa:16:3e:bd:07:7d", "network": {"id": "a013b7aa-023d-4fc6-b0af-978a0806e79f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1099094943-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a9f1052b9944a76b87609da0aa6b8e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap892c26c7-9b", "ovs_interfaceid": "892c26c7-9b46-4a0f-a406-6bdc9c0872fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.160094] env[62521]: DEBUG nova.scheduler.client.report [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.263242] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318584, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.191856} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.263685] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.267498] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cede073-4184-4d65-8ee7-e716824c47eb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.293806] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad/73bc1cf1-7ee1-4a9b-8214-2dc752b700ad.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.296189] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04be3321-5fe3-42d6-8a7d-55733f77f846 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.321332] env[62521]: DEBUG nova.network.neutron [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Successfully updated port: 8824a231-01df-41e5-9bea-9071d8b41cb8 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.333067] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529a40fc-bc8d-3f49-dad1-54db3f0cda4d, 'name': SearchDatastore_Task, 'duration_secs': 0.022817} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.336028] env[62521]: INFO nova.compute.manager [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Took 34.33 seconds to build instance. [ 852.341372] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a92fa1c4-f3ff-4ca9-b39a-47e94e9e9027 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.347688] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 852.347688] env[62521]: value = "task-1318585" [ 852.347688] env[62521]: _type = "Task" [ 852.347688] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.349446] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for the task: (returnval){ [ 852.349446] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5217f838-35e4-301a-962f-59a8d1b5c666" [ 852.349446] env[62521]: _type = "Task" [ 852.349446] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.368395] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318585, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.372982] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5217f838-35e4-301a-962f-59a8d1b5c666, 'name': SearchDatastore_Task, 'duration_secs': 0.017913} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.376583] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.377428] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 2a0ccda4-87a6-4bb3-bf34-4434faf574e6/2a0ccda4-87a6-4bb3-bf34-4434faf574e6.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 852.377428] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7306ca07-f02f-4854-bc73-cde643b9dcc2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.387422] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318583, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.389151] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for the task: (returnval){ [ 852.389151] env[62521]: value = "task-1318586" [ 852.389151] env[62521]: _type = "Task" [ 852.389151] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.397765] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318586, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.478880] env[62521]: DEBUG oslo_concurrency.lockutils [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Releasing lock "refresh_cache-92629ebf-eb50-4862-abe2-01e210e302ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.481300] env[62521]: DEBUG nova.compute.manager [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.482235] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b122ac83-b192-4621-9bec-d63347303f4f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.498749] env[62521]: DEBUG nova.compute.manager [req-f43208dd-d5ea-4d08-87de-e78767d82149 req-68f4ea59-1caa-4024-9fbf-f9904bac944b service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Received event network-changed-8824a231-01df-41e5-9bea-9071d8b41cb8 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.498955] env[62521]: DEBUG nova.compute.manager [req-f43208dd-d5ea-4d08-87de-e78767d82149 req-68f4ea59-1caa-4024-9fbf-f9904bac944b service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Refreshing instance network info cache due to event network-changed-8824a231-01df-41e5-9bea-9071d8b41cb8. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.499212] env[62521]: DEBUG oslo_concurrency.lockutils [req-f43208dd-d5ea-4d08-87de-e78767d82149 req-68f4ea59-1caa-4024-9fbf-f9904bac944b service nova] Acquiring lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.499360] env[62521]: DEBUG oslo_concurrency.lockutils [req-f43208dd-d5ea-4d08-87de-e78767d82149 req-68f4ea59-1caa-4024-9fbf-f9904bac944b service nova] Acquired lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.499547] env[62521]: DEBUG nova.network.neutron [req-f43208dd-d5ea-4d08-87de-e78767d82149 req-68f4ea59-1caa-4024-9fbf-f9904bac944b service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Refreshing network info cache for port 8824a231-01df-41e5-9bea-9071d8b41cb8 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 852.664346] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.336s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.664346] env[62521]: INFO nova.compute.manager [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Migrating [ 852.664346] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.664346] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.665760] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.391s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.666662] env[62521]: DEBUG nova.objects.instance [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lazy-loading 'resources' on Instance uuid b3e33bab-636f-42df-ac6c-667464312c11 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.817785] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.818280] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.818474] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.818699] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.818906] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.822707] env[62521]: INFO nova.compute.manager [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Terminating instance [ 852.826215] env[62521]: DEBUG nova.compute.manager [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 852.826215] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 852.828094] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85912bbc-8637-48d6-b912-a496058451e4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.836388] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.840032] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.840032] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42691b34-b1ca-4780-b780-58b76e4d694b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.840032] env[62521]: DEBUG oslo_concurrency.lockutils [None req-254baab0-e31c-49d0-a865-29204cbad1c5 tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.152s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.845182] env[62521]: DEBUG oslo_vmware.api [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 852.845182] env[62521]: value = "task-1318587" [ 852.845182] env[62521]: _type = "Task" [ 852.845182] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.861383] env[62521]: DEBUG oslo_vmware.api [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.866238] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318585, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.896012] env[62521]: DEBUG oslo_vmware.api [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318583, 'name': PowerOnVM_Task, 'duration_secs': 1.851939} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.899471] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.899471] env[62521]: INFO nova.compute.manager [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Took 9.74 seconds to spawn the instance on the hypervisor. [ 852.899643] env[62521]: DEBUG nova.compute.manager [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.900803] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1ce4b1-67fa-4a3a-a2f3-98768162011b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.912616] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318586, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.090388] env[62521]: DEBUG nova.network.neutron [req-f43208dd-d5ea-4d08-87de-e78767d82149 req-68f4ea59-1caa-4024-9fbf-f9904bac944b service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.168498] env[62521]: INFO nova.compute.rpcapi [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 853.169220] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.296070] env[62521]: DEBUG nova.network.neutron [req-f43208dd-d5ea-4d08-87de-e78767d82149 req-68f4ea59-1caa-4024-9fbf-f9904bac944b service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.368568] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318585, 'name': ReconfigVM_Task, 'duration_secs': 0.537906} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.375414] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad/73bc1cf1-7ee1-4a9b-8214-2dc752b700ad.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.375414] env[62521]: DEBUG oslo_vmware.api [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318587, 'name': PowerOffVM_Task, 'duration_secs': 0.322673} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.378457] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a583fe20-9e71-402d-9631-1f060dbcdaee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.380453] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 853.380640] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 853.381643] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a6817ac-394f-46c5-9387-8e4018898fd9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.387038] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.387038] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.387038] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.387038] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.387038] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.390522] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 853.390522] env[62521]: value = "task-1318588" [ 853.390522] env[62521]: _type = "Task" [ 853.390522] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.391212] env[62521]: INFO nova.compute.manager [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Terminating instance [ 853.396262] env[62521]: DEBUG nova.compute.manager [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 853.396751] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 853.397651] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6608e6-e3ef-4260-aabe-ad0849912f12 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.412937] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318588, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.418397] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.424023] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a98f088-cbc9-4c16-918a-f11f7961fee5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.433319] env[62521]: INFO nova.compute.manager [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Took 30.81 seconds to build instance. [ 853.441254] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "3805d897-fcb5-43ed-b502-7fad71124e74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.441635] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "3805d897-fcb5-43ed-b502-7fad71124e74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.442804] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318586, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.824939} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.444783] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 2a0ccda4-87a6-4bb3-bf34-4434faf574e6/2a0ccda4-87a6-4bb3-bf34-4434faf574e6.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 853.444783] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.445792] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9e5dece2-7d55-4347-ad2c-7cf7b578a5e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.448659] env[62521]: DEBUG oslo_vmware.api [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 853.448659] env[62521]: value = "task-1318590" [ 853.448659] env[62521]: _type = "Task" [ 853.448659] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.451837] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 853.452192] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 853.452471] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleting the datastore file [datastore2] 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.456132] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27abc949-21d5-447e-9056-c3ee3af7fd35 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.464288] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for the task: (returnval){ [ 853.464288] env[62521]: value = "task-1318591" [ 853.464288] env[62521]: _type = "Task" [ 853.464288] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.471834] env[62521]: DEBUG oslo_vmware.api [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 853.471834] env[62521]: value = "task-1318592" [ 853.471834] env[62521]: _type = "Task" [ 853.471834] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.472305] env[62521]: DEBUG oslo_vmware.api [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318590, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.483597] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318591, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.488741] env[62521]: DEBUG oslo_vmware.api [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318592, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.515203] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea9207d-b103-4707-90e1-7dfdc246d926 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.524226] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Doing hard reboot of VM {{(pid=62521) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 853.524667] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-ed10f967-4a94-4052-8e54-8be946334762 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.532708] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9500a49f-b301-4a17-ba18-7b2c443c3f9c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.539257] env[62521]: DEBUG oslo_vmware.api [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 853.539257] env[62521]: value = "task-1318593" [ 853.539257] env[62521]: _type = "Task" [ 853.539257] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.551084] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7317f35c-6f93-40e2-a65b-da47c8a866e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.560954] env[62521]: DEBUG oslo_vmware.api [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318593, 'name': ResetVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.599612] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01d4698-f82b-43dc-b01a-282864c30b8b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.608104] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4e06b8-0cc1-4323-b70b-567e7f17d751 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.623069] env[62521]: DEBUG nova.compute.provider_tree [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 853.692473] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.693214] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.693214] env[62521]: DEBUG nova.network.neutron [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.799774] env[62521]: DEBUG oslo_concurrency.lockutils [req-f43208dd-d5ea-4d08-87de-e78767d82149 req-68f4ea59-1caa-4024-9fbf-f9904bac944b service nova] Releasing lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.799774] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.799986] env[62521]: DEBUG nova.network.neutron [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.906076] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318588, 'name': Rename_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.944590] env[62521]: DEBUG nova.compute.manager [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.948350] env[62521]: DEBUG oslo_concurrency.lockutils [None req-17bea33c-e730-4247-810c-a6cd3584a0c1 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.539s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.969910] env[62521]: DEBUG oslo_vmware.api [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318590, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.986224] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318591, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08336} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.993563] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.994113] env[62521]: DEBUG oslo_vmware.api [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318592, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156811} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.995065] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3891bd-8201-4521-9e81-d3a8f49c789a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.998205] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 853.998444] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 853.998667] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 853.998917] env[62521]: INFO nova.compute.manager [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Took 1.17 seconds to destroy the instance on the hypervisor. [ 853.999248] env[62521]: DEBUG oslo.service.loopingcall [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.999488] env[62521]: DEBUG nova.compute.manager [-] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 853.999611] env[62521]: DEBUG nova.network.neutron [-] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.029441] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 2a0ccda4-87a6-4bb3-bf34-4434faf574e6/2a0ccda4-87a6-4bb3-bf34-4434faf574e6.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.030950] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3caea639-d5df-4f87-a39b-60e065f2874f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.060527] env[62521]: DEBUG oslo_vmware.api [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318593, 'name': ResetVM_Task, 'duration_secs': 0.45115} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.062197] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Did hard reboot of VM {{(pid=62521) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 854.063017] env[62521]: DEBUG nova.compute.manager [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 854.063461] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for the task: (returnval){ [ 854.063461] env[62521]: value = "task-1318594" [ 854.063461] env[62521]: _type = "Task" [ 854.063461] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.064621] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236951a7-e55a-418b-8676-801241fadb85 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.084300] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318594, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.162146] env[62521]: ERROR nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [req-32c8bf1b-e77c-409b-9546-2500f34cd185] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-32c8bf1b-e77c-409b-9546-2500f34cd185"}]} [ 854.182344] env[62521]: DEBUG nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 854.201289] env[62521]: DEBUG nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 854.201682] env[62521]: DEBUG nova.compute.provider_tree [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 854.216328] env[62521]: DEBUG nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 854.241956] env[62521]: DEBUG nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 854.407115] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318588, 'name': Rename_Task, 'duration_secs': 0.625893} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.408064] env[62521]: DEBUG nova.network.neutron [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.410166] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.412996] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e13c828b-6077-4e15-b8b4-0fa5aae20b15 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.423791] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 854.423791] env[62521]: value = "task-1318595" [ 854.423791] env[62521]: _type = "Task" [ 854.423791] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.435315] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318595, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.470763] env[62521]: DEBUG oslo_vmware.api [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318590, 'name': PowerOffVM_Task, 'duration_secs': 0.58804} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.474022] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.474343] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 854.475678] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-444228cf-7d13-4212-87ee-9e7aa73ebee7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.477824] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.536335] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee54c52b-260a-4962-bf9c-27ea02a67c0c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.547473] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f68c1b-d9f9-436c-ab64-73f376f1e392 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.588711] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4badd84a-bf78-42dc-85e8-7cc9379b8d82 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.595251] env[62521]: DEBUG oslo_concurrency.lockutils [None req-43aaaec7-96d3-4e40-b41b-b976c8ffbe19 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "92629ebf-eb50-4862-abe2-01e210e302ad" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.964s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.600439] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318594, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.600978] env[62521]: DEBUG nova.network.neutron [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance_info_cache with network_info: [{"id": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "address": "fa:16:3e:d7:52:1f", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e00f1c5-5c", "ovs_interfaceid": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.608114] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fd6a68-323d-4618-b815-331c0aea6eed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.626103] env[62521]: DEBUG nova.compute.provider_tree [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 854.629957] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 854.630190] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 854.630372] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Deleting the datastore file [datastore2] 583b1a1b-1817-45fc-aa4f-fc9de4b4a243 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.630638] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd8901d4-f32a-4cf9-a5ca-5789d1d14d0a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.636064] env[62521]: DEBUG nova.compute.manager [req-01dda48a-06a7-4a24-9371-436a04f04c43 req-921858e6-241d-406f-b636-f047261a02a5 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Received event network-vif-deleted-a7d4852b-d44e-448f-b25d-d4477b2c3ed9 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.636389] env[62521]: INFO nova.compute.manager [req-01dda48a-06a7-4a24-9371-436a04f04c43 req-921858e6-241d-406f-b636-f047261a02a5 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Neutron deleted interface a7d4852b-d44e-448f-b25d-d4477b2c3ed9; detaching it from the instance and deleting it from the info cache [ 854.636586] env[62521]: DEBUG nova.network.neutron [req-01dda48a-06a7-4a24-9371-436a04f04c43 req-921858e6-241d-406f-b636-f047261a02a5 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.644221] env[62521]: DEBUG oslo_vmware.api [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 854.644221] env[62521]: value = "task-1318597" [ 854.644221] env[62521]: _type = "Task" [ 854.644221] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.655291] env[62521]: DEBUG oslo_vmware.api [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318597, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.791936] env[62521]: DEBUG nova.network.neutron [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [{"id": "8824a231-01df-41e5-9bea-9071d8b41cb8", "address": "fa:16:3e:2e:cc:1b", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8824a231-01", "ovs_interfaceid": "8824a231-01df-41e5-9bea-9071d8b41cb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.870173] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.870481] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.901693] env[62521]: DEBUG nova.network.neutron [-] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.937049] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318595, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.092747] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318594, 'name': ReconfigVM_Task, 'duration_secs': 0.704257} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.093476] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 2a0ccda4-87a6-4bb3-bf34-4434faf574e6/2a0ccda4-87a6-4bb3-bf34-4434faf574e6.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.094370] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-126c3dee-2403-410a-bed7-894718e75757 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.101189] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for the task: (returnval){ [ 855.101189] env[62521]: value = "task-1318598" [ 855.101189] env[62521]: _type = "Task" [ 855.101189] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.106733] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.112175] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318598, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.141502] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3eaa4392-9de4-4824-a07e-5f6fdc281cc4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.150154] env[62521]: ERROR nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [req-7733b87a-f6d3-4f0e-ba27-d4a5112c0d76] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7733b87a-f6d3-4f0e-ba27-d4a5112c0d76"}]} [ 855.153668] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406fff8e-114e-475e-a4c5-9630cfe31f2d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.173489] env[62521]: DEBUG oslo_vmware.api [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318597, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323378} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.173489] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.173489] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 855.173489] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 855.173694] env[62521]: INFO nova.compute.manager [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Took 1.78 seconds to destroy the instance on the hypervisor. [ 855.173918] env[62521]: DEBUG oslo.service.loopingcall [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.174141] env[62521]: DEBUG nova.compute.manager [-] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 855.174236] env[62521]: DEBUG nova.network.neutron [-] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 855.191562] env[62521]: DEBUG nova.compute.manager [req-01dda48a-06a7-4a24-9371-436a04f04c43 req-921858e6-241d-406f-b636-f047261a02a5 service nova] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Detach interface failed, port_id=a7d4852b-d44e-448f-b25d-d4477b2c3ed9, reason: Instance 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 855.192895] env[62521]: DEBUG nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 855.214595] env[62521]: DEBUG nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 855.215090] env[62521]: DEBUG nova.compute.provider_tree [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 855.236748] env[62521]: DEBUG nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 855.266683] env[62521]: DEBUG nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 855.295041] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.295041] env[62521]: DEBUG nova.compute.manager [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Instance network_info: |[{"id": "8824a231-01df-41e5-9bea-9071d8b41cb8", "address": "fa:16:3e:2e:cc:1b", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8824a231-01", "ovs_interfaceid": "8824a231-01df-41e5-9bea-9071d8b41cb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.295041] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:cc:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8824a231-01df-41e5-9bea-9071d8b41cb8', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.304583] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Creating folder: Project (31c788848b0d47478564e53066e7c51a). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.307166] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7d448dd-0b18-43b6-8880-829cd22e9dc0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.319248] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Created folder: Project (31c788848b0d47478564e53066e7c51a) in parent group-v282025. [ 855.319583] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Creating folder: Instances. Parent ref: group-v282101. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.326288] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94df3156-7f7d-4bf0-86e3-b2dcfd1adddf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.338092] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Created folder: Instances in parent group-v282101. [ 855.338270] env[62521]: DEBUG oslo.service.loopingcall [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.338379] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.338640] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97e73485-650d-4f33-9c0d-5b67492d7d06 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.363972] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.363972] env[62521]: value = "task-1318601" [ 855.363972] env[62521]: _type = "Task" [ 855.363972] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.373369] env[62521]: DEBUG nova.compute.manager [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.381120] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318601, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.404694] env[62521]: INFO nova.compute.manager [-] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Took 1.40 seconds to deallocate network for instance. [ 855.439809] env[62521]: DEBUG oslo_vmware.api [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318595, 'name': PowerOnVM_Task, 'duration_secs': 0.676951} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.439809] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.441321] env[62521]: INFO nova.compute.manager [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Took 9.95 seconds to spawn the instance on the hypervisor. [ 855.441321] env[62521]: DEBUG nova.compute.manager [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.441536] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5f04b2-38cd-4381-a3a4-0263643814b3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.620723] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318598, 'name': Rename_Task, 'duration_secs': 0.286556} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.623389] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.623389] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf8f9ce0-b365-4866-ac94-37dd96b7d5f8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.630818] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for the task: (returnval){ [ 855.630818] env[62521]: value = "task-1318602" [ 855.630818] env[62521]: _type = "Task" [ 855.630818] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.640630] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318602, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.643731] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff249388-f86f-4897-8a73-3a9534d64f43 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.652838] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432d0578-21c7-4eed-9c53-cd3d0ed14e6f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.701314] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c2833e-244a-4be7-9928-57cfb86a3928 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.711242] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e23b1b5-f7f0-45a3-a97b-7ca6404fcddd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.729430] env[62521]: DEBUG nova.compute.provider_tree [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 855.829707] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquiring lock "92629ebf-eb50-4862-abe2-01e210e302ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.830284] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "92629ebf-eb50-4862-abe2-01e210e302ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.830391] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquiring lock "92629ebf-eb50-4862-abe2-01e210e302ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.830635] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "92629ebf-eb50-4862-abe2-01e210e302ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.830859] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "92629ebf-eb50-4862-abe2-01e210e302ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.833943] env[62521]: INFO nova.compute.manager [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Terminating instance [ 855.839822] env[62521]: DEBUG nova.compute.manager [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 855.840243] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 855.841786] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb0071d-ef09-4208-a4c1-22d015e6bd04 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.852109] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 855.852475] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce54ba3d-3736-41c0-8b39-6ef6f63e63f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.860248] env[62521]: DEBUG oslo_vmware.api [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 855.860248] env[62521]: value = "task-1318603" [ 855.860248] env[62521]: _type = "Task" [ 855.860248] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.881140] env[62521]: DEBUG oslo_vmware.api [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.887159] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318601, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.914647] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.920237] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.972767] env[62521]: INFO nova.compute.manager [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Took 28.91 seconds to build instance. [ 856.142778] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318602, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.281815] env[62521]: DEBUG nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 79 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 856.282115] env[62521]: DEBUG nova.compute.provider_tree [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 79 to 80 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 856.282309] env[62521]: DEBUG nova.compute.provider_tree [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 856.369984] env[62521]: DEBUG oslo_vmware.api [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318603, 'name': PowerOffVM_Task, 'duration_secs': 0.231837} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.377168] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 856.377168] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 856.377168] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cad6ba3a-2422-43d7-a0b1-b2bc304f88d1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.388751] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318601, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.454968] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 856.455285] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 856.455505] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Deleting the datastore file [datastore2] 92629ebf-eb50-4862-abe2-01e210e302ad {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 856.455806] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b928bdb-48bd-4cbc-8369-a92e3fb2bf45 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.464212] env[62521]: DEBUG oslo_vmware.api [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for the task: (returnval){ [ 856.464212] env[62521]: value = "task-1318605" [ 856.464212] env[62521]: _type = "Task" [ 856.464212] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.476247] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4a8def46-7d35-4a7b-b8ae-41fa9157e6a8 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.854s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.476715] env[62521]: DEBUG oslo_vmware.api [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318605, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.630754] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3e67cd-1b1b-4849-bae5-be7f0c229e81 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.644428] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318602, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.660641] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance 'd5f6492d-ca1b-4e74-b792-b04d55c33660' progress to 0 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 856.674281] env[62521]: DEBUG nova.compute.manager [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received event network-vif-deleted-2ffb5884-fc67-4969-b2e0-c5412152320f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.674578] env[62521]: INFO nova.compute.manager [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Neutron deleted interface 2ffb5884-fc67-4969-b2e0-c5412152320f; detaching it from the instance and deleting it from the info cache [ 856.674832] env[62521]: DEBUG nova.network.neutron [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Updating instance_info_cache with network_info: [{"id": "ef87a463-92f6-449b-8262-f156eaca1596", "address": "fa:16:3e:d7:cc:7e", "network": {"id": "41bc38b6-ff85-44c9-b5b9-1624a2a3368c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2055921925", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.183", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef87a463-92", "ovs_interfaceid": "ef87a463-92f6-449b-8262-f156eaca1596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c6d923d-1406-4e61-943b-a940584e6d6a", "address": "fa:16:3e:82:86:71", "network": {"id": "8218260a-d7e4-469a-bfb9-d5951e9a7e56", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1300759567", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46785c9c-8b22-487d-a854-b3e67c5ed1d7", "external-id": "nsx-vlan-transportzone-430", "segmentation_id": 430, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c6d923d-14", "ovs_interfaceid": "8c6d923d-1406-4e61-943b-a940584e6d6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.789131] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.122s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.790275] env[62521]: DEBUG oslo_concurrency.lockutils [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.452s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.790443] env[62521]: DEBUG nova.objects.instance [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lazy-loading 'resources' on Instance uuid 4afe09db-6c01-444f-a127-6e1f97794544 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.807583] env[62521]: INFO nova.scheduler.client.report [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Deleted allocations for instance b3e33bab-636f-42df-ac6c-667464312c11 [ 856.853017] env[62521]: DEBUG nova.network.neutron [-] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.881763] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318601, 'name': CreateVM_Task, 'duration_secs': 1.433021} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.881763] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 856.883064] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.883064] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.883290] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.883865] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6e2ced1-aa37-4004-b443-1af0169d34db {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.888734] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 856.888734] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c16924-7b04-809d-41f7-db15d9b4e730" [ 856.888734] env[62521]: _type = "Task" [ 856.888734] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.898251] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c16924-7b04-809d-41f7-db15d9b4e730, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.975815] env[62521]: DEBUG oslo_vmware.api [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Task: {'id': task-1318605, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.214912} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.976356] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.976704] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 856.977014] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 856.977234] env[62521]: INFO nova.compute.manager [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Took 1.14 seconds to destroy the instance on the hypervisor. [ 856.977520] env[62521]: DEBUG oslo.service.loopingcall [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.977734] env[62521]: DEBUG nova.compute.manager [-] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 856.977832] env[62521]: DEBUG nova.network.neutron [-] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 857.062577] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "d574f077-b221-4f91-8b54-0915421cb36f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.062933] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "d574f077-b221-4f91-8b54-0915421cb36f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.151479] env[62521]: DEBUG oslo_vmware.api [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318602, 'name': PowerOnVM_Task, 'duration_secs': 1.369639} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.151774] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.151977] env[62521]: INFO nova.compute.manager [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Took 9.21 seconds to spawn the instance on the hypervisor. [ 857.152174] env[62521]: DEBUG nova.compute.manager [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.152969] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c520743-3612-4d6b-8bd7-2591b2db1193 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.166938] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.166938] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6340efe3-dbb6-42e7-862c-b4cfc06c6391 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.172691] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 857.172691] env[62521]: value = "task-1318606" [ 857.172691] env[62521]: _type = "Task" [ 857.172691] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.180859] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66e5aac0-aac0-446d-8529-73e9661575c8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.182696] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318606, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.189622] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0da0a5-fb4e-4750-bb74-8976dc5e9a7e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.230556] env[62521]: DEBUG nova.compute.manager [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Detach interface failed, port_id=2ffb5884-fc67-4969-b2e0-c5412152320f, reason: Instance 583b1a1b-1817-45fc-aa4f-fc9de4b4a243 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 857.230556] env[62521]: DEBUG nova.compute.manager [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received event network-vif-deleted-8c6d923d-1406-4e61-943b-a940584e6d6a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.230767] env[62521]: INFO nova.compute.manager [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Neutron deleted interface 8c6d923d-1406-4e61-943b-a940584e6d6a; detaching it from the instance and deleting it from the info cache [ 857.231110] env[62521]: DEBUG nova.network.neutron [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Updating instance_info_cache with network_info: [{"id": "ef87a463-92f6-449b-8262-f156eaca1596", "address": "fa:16:3e:d7:cc:7e", "network": {"id": "41bc38b6-ff85-44c9-b5b9-1624a2a3368c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2055921925", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.183", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef87a463-92", "ovs_interfaceid": "ef87a463-92f6-449b-8262-f156eaca1596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.317382] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0bfa1666-8733-49f6-a298-4dc6482989ce tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "b3e33bab-636f-42df-ac6c-667464312c11" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.123s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.358134] env[62521]: INFO nova.compute.manager [-] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Took 2.18 seconds to deallocate network for instance. [ 857.410204] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c16924-7b04-809d-41f7-db15d9b4e730, 'name': SearchDatastore_Task, 'duration_secs': 0.017255} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.410204] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.410204] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.410204] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.410204] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.410204] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.410204] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4edbe088-8e19-4c02-8463-97dd804dabdc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.435052] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.435052] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 857.439302] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3fc1b20-a8ff-473b-a805-fff0659486cf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.444642] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 857.444642] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526a7f9c-4346-8aaa-6274-61277d18079b" [ 857.444642] env[62521]: _type = "Task" [ 857.444642] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.458558] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526a7f9c-4346-8aaa-6274-61277d18079b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.566763] env[62521]: DEBUG nova.compute.manager [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 857.623025] env[62521]: INFO nova.compute.manager [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Rebuilding instance [ 857.635474] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b07b4b92-5463-4d53-8f46-75a57f83810b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.646344] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4449aa-6a1a-4a78-99d3-bf02eb41f2dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.694564] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa2d8ac-1d5f-4e68-87b7-feb188d70c9f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.701694] env[62521]: INFO nova.compute.manager [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Took 28.77 seconds to build instance. [ 857.706786] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318606, 'name': PowerOffVM_Task, 'duration_secs': 0.290456} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.714067] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 857.714307] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance 'd5f6492d-ca1b-4e74-b792-b04d55c33660' progress to 17 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 857.721144] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42015e4-acf6-4ab4-941e-2902b6a9450b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.739176] env[62521]: DEBUG nova.compute.provider_tree [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 857.743501] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30ac7d24-6997-45ec-abf1-9e372d4b9b90 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.753166] env[62521]: DEBUG nova.compute.manager [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.754722] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55353328-bc4f-4487-97de-a587b3d63ccc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.763055] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d057e0-dd25-40a2-a660-773ed68f7be0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.776656] env[62521]: DEBUG nova.compute.manager [req-b343a943-1082-4e7f-a146-037d8456adcf req-6f7ae770-6ac7-45ab-b7a7-6dab448b985c service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received event network-changed-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.776896] env[62521]: DEBUG nova.compute.manager [req-b343a943-1082-4e7f-a146-037d8456adcf req-6f7ae770-6ac7-45ab-b7a7-6dab448b985c service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Refreshing instance network info cache due to event network-changed-72671c1a-eabd-4a80-a591-17b26cea0249. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.777175] env[62521]: DEBUG oslo_concurrency.lockutils [req-b343a943-1082-4e7f-a146-037d8456adcf req-6f7ae770-6ac7-45ab-b7a7-6dab448b985c service nova] Acquiring lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.777364] env[62521]: DEBUG oslo_concurrency.lockutils [req-b343a943-1082-4e7f-a146-037d8456adcf req-6f7ae770-6ac7-45ab-b7a7-6dab448b985c service nova] Acquired lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.777559] env[62521]: DEBUG nova.network.neutron [req-b343a943-1082-4e7f-a146-037d8456adcf req-6f7ae770-6ac7-45ab-b7a7-6dab448b985c service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Refreshing network info cache for port 72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 857.800729] env[62521]: DEBUG nova.compute.manager [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Detach interface failed, port_id=8c6d923d-1406-4e61-943b-a940584e6d6a, reason: Instance 583b1a1b-1817-45fc-aa4f-fc9de4b4a243 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 857.802569] env[62521]: DEBUG nova.compute.manager [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Received event network-vif-deleted-ef87a463-92f6-449b-8262-f156eaca1596 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.802569] env[62521]: INFO nova.compute.manager [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Neutron deleted interface ef87a463-92f6-449b-8262-f156eaca1596; detaching it from the instance and deleting it from the info cache [ 857.802569] env[62521]: DEBUG nova.network.neutron [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.868149] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.957981] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526a7f9c-4346-8aaa-6274-61277d18079b, 'name': SearchDatastore_Task, 'duration_secs': 0.014821} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.958921] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c2b82f0-ec12-4bb9-8a2b-16dcffa15da7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.964862] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 857.964862] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a9461b-8e9b-a935-62a8-86c6e22844ce" [ 857.964862] env[62521]: _type = "Task" [ 857.964862] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.973127] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a9461b-8e9b-a935-62a8-86c6e22844ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.106998] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.204144] env[62521]: DEBUG oslo_concurrency.lockutils [None req-990c28da-d18b-47da-bb61-56854024b4ef tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.825s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.223350] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.223564] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.224265] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.224265] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.224265] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.224265] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.224502] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.224533] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.224995] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.224995] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.225135] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.230901] env[62521]: DEBUG nova.network.neutron [-] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.232185] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4535a72e-77a3-4a18-a960-a89b1c612d44 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.251187] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 858.251187] env[62521]: value = "task-1318607" [ 858.251187] env[62521]: _type = "Task" [ 858.251187] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.261198] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318607, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.272514] env[62521]: ERROR nova.scheduler.client.report [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] [req-4c741cf0-13e4-484f-8c54-4a712660f892] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4c741cf0-13e4-484f-8c54-4a712660f892"}]} [ 858.286875] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 858.287353] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15864ba2-82b7-4acd-b59a-65a376be1162 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.297284] env[62521]: DEBUG nova.scheduler.client.report [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 858.299467] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for the task: (returnval){ [ 858.299467] env[62521]: value = "task-1318608" [ 858.299467] env[62521]: _type = "Task" [ 858.299467] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.305598] env[62521]: DEBUG oslo_vmware.rw_handles [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec15df-88a4-bf8a-634b-31965c48977a/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 858.307239] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023bab1f-6957-4fcd-ac79-8073c9585d42 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.311018] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59396ae1-e235-4c70-bd2d-a964507f6b38 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.317148] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318608, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.318377] env[62521]: DEBUG nova.scheduler.client.report [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 858.318629] env[62521]: DEBUG nova.compute.provider_tree [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 858.322765] env[62521]: DEBUG oslo_vmware.rw_handles [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec15df-88a4-bf8a-634b-31965c48977a/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 858.322851] env[62521]: ERROR oslo_vmware.rw_handles [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec15df-88a4-bf8a-634b-31965c48977a/disk-0.vmdk due to incomplete transfer. [ 858.323113] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-00ab1e50-2698-41d5-af2c-966cfbdae126 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.331079] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d201ef-2d1c-4469-bcbd-974f4c786494 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.342141] env[62521]: DEBUG nova.scheduler.client.report [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 858.345533] env[62521]: DEBUG oslo_vmware.rw_handles [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec15df-88a4-bf8a-634b-31965c48977a/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 858.345617] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Uploaded image 9f64b520-b041-469a-8080-6b597c951678 to the Glance image server {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 858.348103] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 858.351980] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1a646b98-51d5-4bb9-99ef-bc4232eb8acd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.359106] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 858.359106] env[62521]: value = "task-1318609" [ 858.359106] env[62521]: _type = "Task" [ 858.359106] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.371142] env[62521]: DEBUG nova.compute.manager [req-8abaef71-6469-415d-95b6-82cdc53f5de2 req-436fd39e-e780-4ed9-a4ba-968d7078d303 service nova] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Detach interface failed, port_id=ef87a463-92f6-449b-8262-f156eaca1596, reason: Instance 583b1a1b-1817-45fc-aa4f-fc9de4b4a243 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 858.377714] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318609, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.378871] env[62521]: DEBUG nova.scheduler.client.report [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 858.480629] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a9461b-8e9b-a935-62a8-86c6e22844ce, 'name': SearchDatastore_Task, 'duration_secs': 0.013297} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.481020] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.481376] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 7b5afeab-f04e-453a-a3ed-09e3a725e8d6/7b5afeab-f04e-453a-a3ed-09e3a725e8d6.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 858.484820] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79994c5c-1963-426b-b36b-3ddeb0985a8c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.492677] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 858.492677] env[62521]: value = "task-1318610" [ 858.492677] env[62521]: _type = "Task" [ 858.492677] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.508835] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318610, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.676888] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbcfd659-984d-4b4c-8574-e44d946d2a79 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.684358] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed651f4b-0c4b-498e-b665-14d1ab12b492 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.717717] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafee9ed-a998-4ff0-8850-72ea8bea19c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.721367] env[62521]: DEBUG nova.compute.manager [req-dc9d9a4f-1f2d-4d9d-932a-b9f46bd26c39 req-e405b5f4-eea9-43e4-999a-8308e7b76e98 service nova] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Received event network-vif-deleted-892c26c7-9b46-4a0f-a406-6bdc9c0872fa {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.727466] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87ee4c9-d146-4f0c-8745-5ceddb4f892b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.743225] env[62521]: DEBUG nova.compute.provider_tree [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 858.744567] env[62521]: INFO nova.compute.manager [-] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Took 1.77 seconds to deallocate network for instance. [ 858.761178] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318607, 'name': ReconfigVM_Task, 'duration_secs': 0.238841} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.761554] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance 'd5f6492d-ca1b-4e74-b792-b04d55c33660' progress to 33 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 858.809267] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318608, 'name': PowerOffVM_Task, 'duration_secs': 0.296489} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.809545] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.810368] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 858.810540] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7451a48-9e01-453c-a2a4-03ce33194d6a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.816740] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for the task: (returnval){ [ 858.816740] env[62521]: value = "task-1318611" [ 858.816740] env[62521]: _type = "Task" [ 858.816740] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.827781] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 858.828112] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 858.828244] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282049', 'volume_id': '683a2b16-e997-4b24-90c9-559ca2510a58', 'name': 'volume-683a2b16-e997-4b24-90c9-559ca2510a58', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8cabcdc-d728-434d-a40d-fe751db086c3', 'attached_at': '', 'detached_at': '', 'volume_id': '683a2b16-e997-4b24-90c9-559ca2510a58', 'serial': '683a2b16-e997-4b24-90c9-559ca2510a58'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 858.829086] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9443df6f-484a-4ef6-9c49-02106a4d7841 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.848363] env[62521]: DEBUG nova.network.neutron [req-b343a943-1082-4e7f-a146-037d8456adcf req-6f7ae770-6ac7-45ab-b7a7-6dab448b985c service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updated VIF entry in instance network info cache for port 72671c1a-eabd-4a80-a591-17b26cea0249. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 858.848741] env[62521]: DEBUG nova.network.neutron [req-b343a943-1082-4e7f-a146-037d8456adcf req-6f7ae770-6ac7-45ab-b7a7-6dab448b985c service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating instance_info_cache with network_info: [{"id": "72671c1a-eabd-4a80-a591-17b26cea0249", "address": "fa:16:3e:3a:1f:96", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72671c1a-ea", "ovs_interfaceid": "72671c1a-eabd-4a80-a591-17b26cea0249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.850422] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e4ca26-51e7-4124-89f2-71c3614e9272 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.858173] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e747f34-42f5-4fe1-acf8-d3284811d40f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.884309] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318609, 'name': Destroy_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.885106] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbe9bf3-c188-419f-8a9d-a1d24e2fb17a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.900895] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] The volume has not been displaced from its original location: [datastore2] volume-683a2b16-e997-4b24-90c9-559ca2510a58/volume-683a2b16-e997-4b24-90c9-559ca2510a58.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 858.906598] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Reconfiguring VM instance instance-00000038 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 858.906986] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c12ee67a-22ac-43ba-8dce-818af19aaf00 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.927019] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for the task: (returnval){ [ 858.927019] env[62521]: value = "task-1318612" [ 858.927019] env[62521]: _type = "Task" [ 858.927019] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.937794] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318612, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.002538] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318610, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.252697] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.268767] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.268767] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.268767] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.269090] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.269090] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.269271] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.269545] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.269713] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.269889] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.270697] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.270914] env[62521]: DEBUG nova.virt.hardware [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.280084] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Reconfiguring VM instance instance-0000003b to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 859.280436] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c620c75f-8f36-47d6-a6da-73ec86e22b08 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.294403] env[62521]: DEBUG nova.scheduler.client.report [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 81 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 859.294664] env[62521]: DEBUG nova.compute.provider_tree [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 81 to 82 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 859.294850] env[62521]: DEBUG nova.compute.provider_tree [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 859.304663] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 859.304663] env[62521]: value = "task-1318613" [ 859.304663] env[62521]: _type = "Task" [ 859.304663] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.315313] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318613, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.355213] env[62521]: DEBUG oslo_concurrency.lockutils [req-b343a943-1082-4e7f-a146-037d8456adcf req-6f7ae770-6ac7-45ab-b7a7-6dab448b985c service nova] Releasing lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.373525] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318609, 'name': Destroy_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.437498] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318612, 'name': ReconfigVM_Task, 'duration_secs': 0.25424} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.437762] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Reconfigured VM instance instance-00000038 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 859.443097] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-859d8375-2cd1-4104-b655-20cb025bcd8f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.460229] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for the task: (returnval){ [ 859.460229] env[62521]: value = "task-1318614" [ 859.460229] env[62521]: _type = "Task" [ 859.460229] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.468906] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318614, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.503137] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318610, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.940798} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.503608] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 7b5afeab-f04e-453a-a3ed-09e3a725e8d6/7b5afeab-f04e-453a-a3ed-09e3a725e8d6.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.503752] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.504489] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-efe62117-0efe-4653-8e4f-2cf41dffedc6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.511848] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 859.511848] env[62521]: value = "task-1318615" [ 859.511848] env[62521]: _type = "Task" [ 859.511848] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.520935] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318615, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.684674] env[62521]: DEBUG oslo_concurrency.lockutils [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.684934] env[62521]: DEBUG oslo_concurrency.lockutils [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.801024] env[62521]: DEBUG oslo_concurrency.lockutils [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.011s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.804016] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.095s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.804326] env[62521]: DEBUG nova.objects.instance [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62521) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 859.819920] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318613, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.820871] env[62521]: INFO nova.scheduler.client.report [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Deleted allocations for instance 4afe09db-6c01-444f-a127-6e1f97794544 [ 859.871381] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318609, 'name': Destroy_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.970794] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318614, 'name': ReconfigVM_Task, 'duration_secs': 0.18323} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.971201] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282049', 'volume_id': '683a2b16-e997-4b24-90c9-559ca2510a58', 'name': 'volume-683a2b16-e997-4b24-90c9-559ca2510a58', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8cabcdc-d728-434d-a40d-fe751db086c3', 'attached_at': '', 'detached_at': '', 'volume_id': '683a2b16-e997-4b24-90c9-559ca2510a58', 'serial': '683a2b16-e997-4b24-90c9-559ca2510a58'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 859.971520] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.972360] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036c0963-8286-42ce-8cb3-447e7474f1c3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.979105] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 859.979366] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c3d023f-ab2c-4d25-aa33-c9eb6e45a1a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.022807] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318615, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071177} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.022807] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.025228] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9a205e-522b-4707-b7af-fada07d11d17 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.055275] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 7b5afeab-f04e-453a-a3ed-09e3a725e8d6/7b5afeab-f04e-453a-a3ed-09e3a725e8d6.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.057589] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db1985c7-cc5e-492e-82b2-5aab38fb0aad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.071081] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.071315] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.072028] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Deleting the datastore file [datastore2] e8cabcdc-d728-434d-a40d-fe751db086c3 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.072028] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5055d6c6-a9a6-4ad2-b1b1-e39cdec1eddd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.079087] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 860.079087] env[62521]: value = "task-1318618" [ 860.079087] env[62521]: _type = "Task" [ 860.079087] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.079964] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for the task: (returnval){ [ 860.079964] env[62521]: value = "task-1318617" [ 860.079964] env[62521]: _type = "Task" [ 860.079964] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.093071] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318617, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.096129] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318618, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.189284] env[62521]: DEBUG nova.compute.utils [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.198147] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "1d1d2235-748a-4f27-a915-b69f9110257b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.198664] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "1d1d2235-748a-4f27-a915-b69f9110257b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.266814] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "3447712d-daa6-4329-8015-6474cca1a5e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.267104] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "3447712d-daa6-4329-8015-6474cca1a5e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.321026] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318613, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.327641] env[62521]: DEBUG oslo_concurrency.lockutils [None req-525630d0-1546-4558-8526-145c48444782 tempest-InstanceActionsNegativeTestJSON-555331121 tempest-InstanceActionsNegativeTestJSON-555331121-project-member] Lock "4afe09db-6c01-444f-a127-6e1f97794544" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.966s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.371464] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318609, 'name': Destroy_Task, 'duration_secs': 1.821182} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.371735] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Destroyed the VM [ 860.372067] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 860.372617] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3639076c-ff6d-471d-9c27-c3d3c2c32932 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.378534] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 860.378534] env[62521]: value = "task-1318619" [ 860.378534] env[62521]: _type = "Task" [ 860.378534] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.386527] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318619, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.594290] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318618, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.597587] env[62521]: DEBUG oslo_vmware.api [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Task: {'id': task-1318617, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1114} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.597837] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.598043] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.598266] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.658179] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 860.658527] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86cb66e5-62d7-4d8d-a6e3-b6f226368367 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.667533] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774fff44-35b5-4110-8a1c-fdc810683bee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.696290] env[62521]: DEBUG oslo_concurrency.lockutils [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.698047] env[62521]: ERROR nova.compute.manager [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Failed to detach volume 683a2b16-e997-4b24-90c9-559ca2510a58 from /dev/sda: nova.exception.InstanceNotFound: Instance e8cabcdc-d728-434d-a40d-fe751db086c3 could not be found. [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Traceback (most recent call last): [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self.driver.rebuild(**kwargs) [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] raise NotImplementedError() [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] NotImplementedError [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] During handling of the above exception, another exception occurred: [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Traceback (most recent call last): [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self.driver.detach_volume(context, old_connection_info, [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] return self._volumeops.detach_volume(connection_info, instance) [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self._detach_volume_vmdk(connection_info, instance) [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] stable_ref.fetch_moref(session) [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] raise exception.InstanceNotFound(instance_id=self._uuid) [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] nova.exception.InstanceNotFound: Instance e8cabcdc-d728-434d-a40d-fe751db086c3 could not be found. [ 860.698047] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] [ 860.701400] env[62521]: DEBUG nova.compute.manager [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 860.769426] env[62521]: DEBUG nova.compute.manager [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 860.819603] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d5739100-8094-42ae-8265-e7d526a73328 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.820818] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318613, 'name': ReconfigVM_Task, 'duration_secs': 1.300454} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.821771] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.343s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.822683] env[62521]: INFO nova.compute.claims [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 860.826173] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Reconfigured VM instance instance-0000003b to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 860.827647] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ecdd892-1fc4-4f06-bcc1-fc344c0928a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.857516] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] d5f6492d-ca1b-4e74-b792-b04d55c33660/d5f6492d-ca1b-4e74-b792-b04d55c33660.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.859553] env[62521]: DEBUG nova.compute.utils [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Build of instance e8cabcdc-d728-434d-a40d-fe751db086c3 aborted: Failed to rebuild volume backed instance. {{(pid=62521) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 860.862293] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba003cfc-121f-4a8c-ae48-9ba6c54db43f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.879089] env[62521]: ERROR nova.compute.manager [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance e8cabcdc-d728-434d-a40d-fe751db086c3 aborted: Failed to rebuild volume backed instance. [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Traceback (most recent call last): [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self.driver.rebuild(**kwargs) [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] raise NotImplementedError() [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] NotImplementedError [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] During handling of the above exception, another exception occurred: [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Traceback (most recent call last): [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self._detach_root_volume(context, instance, root_bdm) [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] with excutils.save_and_reraise_exception(): [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self.force_reraise() [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] raise self.value [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self.driver.detach_volume(context, old_connection_info, [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] return self._volumeops.detach_volume(connection_info, instance) [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self._detach_volume_vmdk(connection_info, instance) [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] stable_ref.fetch_moref(session) [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] raise exception.InstanceNotFound(instance_id=self._uuid) [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] nova.exception.InstanceNotFound: Instance e8cabcdc-d728-434d-a40d-fe751db086c3 could not be found. [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] During handling of the above exception, another exception occurred: [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Traceback (most recent call last): [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] yield [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 860.879089] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self._do_rebuild_instance_with_claim( [ 860.880259] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 860.880259] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self._do_rebuild_instance( [ 860.880259] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 860.880259] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self._rebuild_default_impl(**kwargs) [ 860.880259] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 860.880259] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] self._rebuild_volume_backed_instance( [ 860.880259] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 860.880259] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] raise exception.BuildAbortException( [ 860.880259] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] nova.exception.BuildAbortException: Build of instance e8cabcdc-d728-434d-a40d-fe751db086c3 aborted: Failed to rebuild volume backed instance. [ 860.880259] env[62521]: ERROR nova.compute.manager [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] [ 860.888210] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 860.888210] env[62521]: value = "task-1318620" [ 860.888210] env[62521]: _type = "Task" [ 860.888210] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.895813] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318619, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.902709] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318620, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.013166] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquiring lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.013619] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.013833] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquiring lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.014119] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.014335] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.017218] env[62521]: INFO nova.compute.manager [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Terminating instance [ 861.019281] env[62521]: DEBUG nova.compute.manager [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 861.019488] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 861.020494] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8570251a-c9cf-4e15-bdcd-5816c7c40a79 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.028524] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 861.028814] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a50c9c5a-c2dd-4e48-8455-0f6dba9eaa08 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.035737] env[62521]: DEBUG oslo_vmware.api [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for the task: (returnval){ [ 861.035737] env[62521]: value = "task-1318621" [ 861.035737] env[62521]: _type = "Task" [ 861.035737] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.045298] env[62521]: DEBUG oslo_vmware.api [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318621, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.090874] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318618, 'name': ReconfigVM_Task, 'duration_secs': 0.714495} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.091228] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 7b5afeab-f04e-453a-a3ed-09e3a725e8d6/7b5afeab-f04e-453a-a3ed-09e3a725e8d6.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.091835] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9453a7bd-f87d-4bb1-ba24-72dbbcce0b3d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.097887] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 861.097887] env[62521]: value = "task-1318622" [ 861.097887] env[62521]: _type = "Task" [ 861.097887] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.106206] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318622, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.236697] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.294999] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.402092] env[62521]: DEBUG oslo_vmware.api [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318619, 'name': RemoveSnapshot_Task, 'duration_secs': 0.92345} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.403731] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 861.404064] env[62521]: INFO nova.compute.manager [None req-219b18c7-71b7-42f7-a1ac-a837708fd77d tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Took 17.91 seconds to snapshot the instance on the hypervisor. [ 861.406466] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318620, 'name': ReconfigVM_Task, 'duration_secs': 0.378605} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.407013] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Reconfigured VM instance instance-0000003b to attach disk [datastore1] d5f6492d-ca1b-4e74-b792-b04d55c33660/d5f6492d-ca1b-4e74-b792-b04d55c33660.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.407315] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance 'd5f6492d-ca1b-4e74-b792-b04d55c33660' progress to 50 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 861.547916] env[62521]: DEBUG oslo_vmware.api [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318621, 'name': PowerOffVM_Task, 'duration_secs': 0.232852} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.548389] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 861.548699] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 861.549487] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97526494-281d-4e83-950c-f451a88e3218 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.609821] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318622, 'name': Rename_Task, 'duration_secs': 0.185355} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.609821] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.609821] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-278f012a-9e63-404e-b6b2-a2bf0ab6b2bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.615148] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 861.615148] env[62521]: value = "task-1318624" [ 861.615148] env[62521]: _type = "Task" [ 861.615148] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.621382] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318624, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.662563] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 861.662563] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 861.662563] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Deleting the datastore file [datastore1] 2a0ccda4-87a6-4bb3-bf34-4434faf574e6 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 861.662563] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa7aa4a8-6ac1-438b-8d60-9ef1cc10d5be {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.673288] env[62521]: DEBUG oslo_vmware.api [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for the task: (returnval){ [ 861.673288] env[62521]: value = "task-1318625" [ 861.673288] env[62521]: _type = "Task" [ 861.673288] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.679524] env[62521]: DEBUG oslo_vmware.api [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318625, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.801511] env[62521]: DEBUG oslo_concurrency.lockutils [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.801511] env[62521]: DEBUG oslo_concurrency.lockutils [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.801511] env[62521]: INFO nova.compute.manager [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Attaching volume 6e8eee93-7abc-4c71-8667-16b6b4dbd033 to /dev/sdb [ 861.836490] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc4050d-edd1-4b1b-b883-3e1a9b43f0cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.853791] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc86dc6-61ab-41eb-9014-edaea9b6fa01 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.874515] env[62521]: DEBUG nova.virt.block_device [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updating existing volume attachment record: 31047706-c1cf-4d61-96ec-6eae07a5a2dd {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 861.923914] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a99e19-55b7-49f6-a1cf-a4ae7480e836 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.949831] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9acce6a-e505-4d76-a549-d825a0a4ba90 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.969936] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance 'd5f6492d-ca1b-4e74-b792-b04d55c33660' progress to 67 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 862.129059] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318624, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.163997] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9f0069-97a0-4a70-afae-45a00c6f2af2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.186037] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2a375a-6eeb-44bb-898e-ea4b4a458bb0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.186037] env[62521]: DEBUG oslo_vmware.api [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Task: {'id': task-1318625, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176132} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.186037] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 862.186037] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 862.186296] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 862.186350] env[62521]: INFO nova.compute.manager [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Took 1.17 seconds to destroy the instance on the hypervisor. [ 862.187033] env[62521]: DEBUG oslo.service.loopingcall [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.187033] env[62521]: DEBUG nova.compute.manager [-] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 862.187033] env[62521]: DEBUG nova.network.neutron [-] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 862.223019] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80879682-1154-48c5-894c-ef22d1be6b8e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.228367] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2ebfd0-2435-49b7-9c39-39018f34df47 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.243802] env[62521]: DEBUG nova.compute.provider_tree [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.535520] env[62521]: DEBUG nova.network.neutron [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Port 7e00f1c5-5c15-430f-baff-cf464bf5d9b3 binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 862.565680] env[62521]: DEBUG nova.compute.manager [req-468944eb-ddc5-49d7-b443-49a17e01f6ca req-c5e798ef-a8d9-48ca-a49d-24efe88996d9 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Received event network-vif-deleted-20597486-a59c-4869-9c6e-3d10e7b83d5f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.566046] env[62521]: INFO nova.compute.manager [req-468944eb-ddc5-49d7-b443-49a17e01f6ca req-c5e798ef-a8d9-48ca-a49d-24efe88996d9 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Neutron deleted interface 20597486-a59c-4869-9c6e-3d10e7b83d5f; detaching it from the instance and deleting it from the info cache [ 862.566347] env[62521]: DEBUG nova.network.neutron [req-468944eb-ddc5-49d7-b443-49a17e01f6ca req-c5e798ef-a8d9-48ca-a49d-24efe88996d9 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.625892] env[62521]: DEBUG oslo_vmware.api [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318624, 'name': PowerOnVM_Task, 'duration_secs': 0.953989} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.626328] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.626416] env[62521]: INFO nova.compute.manager [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Took 12.23 seconds to spawn the instance on the hypervisor. [ 862.626606] env[62521]: DEBUG nova.compute.manager [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.627449] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0351f00f-1678-4df8-9744-d1b9e53fc285 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.751536] env[62521]: DEBUG nova.scheduler.client.report [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.901043] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.998290] env[62521]: DEBUG nova.network.neutron [-] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.072745] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de9d9e13-cda0-478b-8165-237858046094 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.083021] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6074a2-a4bd-4065-b3a3-3266064c20a4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.111526] env[62521]: DEBUG nova.compute.manager [req-468944eb-ddc5-49d7-b443-49a17e01f6ca req-c5e798ef-a8d9-48ca-a49d-24efe88996d9 service nova] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Detach interface failed, port_id=20597486-a59c-4869-9c6e-3d10e7b83d5f, reason: Instance 2a0ccda4-87a6-4bb3-bf34-4434faf574e6 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 863.145262] env[62521]: INFO nova.compute.manager [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Took 29.18 seconds to build instance. [ 863.259030] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.259030] env[62521]: DEBUG nova.compute.manager [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 863.261662] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.347s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.263781] env[62521]: INFO nova.compute.claims [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.357598] env[62521]: DEBUG oslo_concurrency.lockutils [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Acquiring lock "e8cabcdc-d728-434d-a40d-fe751db086c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.357598] env[62521]: DEBUG oslo_concurrency.lockutils [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "e8cabcdc-d728-434d-a40d-fe751db086c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.357750] env[62521]: DEBUG oslo_concurrency.lockutils [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Acquiring lock "e8cabcdc-d728-434d-a40d-fe751db086c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.357827] env[62521]: DEBUG oslo_concurrency.lockutils [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "e8cabcdc-d728-434d-a40d-fe751db086c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.358036] env[62521]: DEBUG oslo_concurrency.lockutils [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "e8cabcdc-d728-434d-a40d-fe751db086c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.360761] env[62521]: INFO nova.compute.manager [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Terminating instance [ 863.362717] env[62521]: DEBUG nova.compute.manager [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 863.363094] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa79b33e-153d-421d-9115-e11634941419 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.372465] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175fbc96-2aa3-4591-93f9-9cd3325ff4ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.398751] env[62521]: WARNING nova.virt.vmwareapi.driver [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance e8cabcdc-d728-434d-a40d-fe751db086c3 could not be found. [ 863.398751] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 863.398983] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4165eae7-3598-47af-84e4-410314a52a39 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.406880] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bccbac-d562-42fb-a873-8db37979d4ae {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.431869] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e8cabcdc-d728-434d-a40d-fe751db086c3 could not be found. [ 863.432089] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 863.432277] env[62521]: INFO nova.compute.manager [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Took 0.07 seconds to destroy the instance on the hypervisor. [ 863.432524] env[62521]: DEBUG oslo.service.loopingcall [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.432749] env[62521]: DEBUG nova.compute.manager [-] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 863.432846] env[62521]: DEBUG nova.network.neutron [-] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 863.502265] env[62521]: INFO nova.compute.manager [-] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Took 1.31 seconds to deallocate network for instance. [ 863.559715] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "d5f6492d-ca1b-4e74-b792-b04d55c33660-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.559715] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.559715] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.647173] env[62521]: DEBUG oslo_concurrency.lockutils [None req-205bdb18-e57c-4b8e-9a68-d8d209fe68e7 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.522s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.769438] env[62521]: DEBUG nova.compute.utils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 863.776881] env[62521]: DEBUG nova.compute.manager [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 863.776881] env[62521]: DEBUG nova.network.neutron [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 863.826540] env[62521]: DEBUG nova.policy [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d7aab6f3f6245c2b56979677428a9df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e6baab1556e4b00b07ab554664b4445', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.007854] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.132180] env[62521]: DEBUG nova.network.neutron [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Successfully created port: 1643da59-fa10-4645-994b-0c557c3a6e10 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 864.158421] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "e2104c11-7713-4542-83da-bc4d2534113a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.158730] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "e2104c11-7713-4542-83da-bc4d2534113a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.257814] env[62521]: DEBUG nova.compute.manager [req-c4f12dc2-fc3b-43a7-9ca9-6b65b9b6b1db req-926f0e60-fbd1-4c9d-b35a-5ba5a8de1526 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Received event network-changed-8824a231-01df-41e5-9bea-9071d8b41cb8 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.258064] env[62521]: DEBUG nova.compute.manager [req-c4f12dc2-fc3b-43a7-9ca9-6b65b9b6b1db req-926f0e60-fbd1-4c9d-b35a-5ba5a8de1526 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Refreshing instance network info cache due to event network-changed-8824a231-01df-41e5-9bea-9071d8b41cb8. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 864.258227] env[62521]: DEBUG oslo_concurrency.lockutils [req-c4f12dc2-fc3b-43a7-9ca9-6b65b9b6b1db req-926f0e60-fbd1-4c9d-b35a-5ba5a8de1526 service nova] Acquiring lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.258405] env[62521]: DEBUG oslo_concurrency.lockutils [req-c4f12dc2-fc3b-43a7-9ca9-6b65b9b6b1db req-926f0e60-fbd1-4c9d-b35a-5ba5a8de1526 service nova] Acquired lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.258738] env[62521]: DEBUG nova.network.neutron [req-c4f12dc2-fc3b-43a7-9ca9-6b65b9b6b1db req-926f0e60-fbd1-4c9d-b35a-5ba5a8de1526 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Refreshing network info cache for port 8824a231-01df-41e5-9bea-9071d8b41cb8 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 864.277988] env[62521]: DEBUG nova.compute.manager [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 864.606283] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a50cf1-f8f4-439e-b447-8c2aea6895e2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.614281] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95a50ed-8d1e-42a9-b485-30d1f47ef2d3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.646963] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.647207] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.647411] env[62521]: DEBUG nova.network.neutron [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 864.649375] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-690f172e-ec1d-4b13-8b7f-aec6cf980a16 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.658851] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb172121-6e5a-40f5-a39b-cca82016a5bc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.664484] env[62521]: DEBUG nova.compute.manager [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.677785] env[62521]: DEBUG nova.compute.provider_tree [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.773659] env[62521]: DEBUG nova.network.neutron [-] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.181480] env[62521]: DEBUG nova.scheduler.client.report [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.190847] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.276090] env[62521]: INFO nova.compute.manager [-] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Took 1.84 seconds to deallocate network for instance. [ 865.287563] env[62521]: DEBUG nova.compute.manager [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 865.330123] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 865.330388] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 865.330747] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.330977] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 865.331155] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.331391] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 865.331704] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 865.331854] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 865.332041] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 865.332218] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 865.332397] env[62521]: DEBUG nova.virt.hardware [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.334455] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd3453c-3ae8-4e52-bf3f-012a59e1d5fb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.340240] env[62521]: DEBUG nova.network.neutron [req-c4f12dc2-fc3b-43a7-9ca9-6b65b9b6b1db req-926f0e60-fbd1-4c9d-b35a-5ba5a8de1526 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updated VIF entry in instance network info cache for port 8824a231-01df-41e5-9bea-9071d8b41cb8. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 865.340580] env[62521]: DEBUG nova.network.neutron [req-c4f12dc2-fc3b-43a7-9ca9-6b65b9b6b1db req-926f0e60-fbd1-4c9d-b35a-5ba5a8de1526 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [{"id": "8824a231-01df-41e5-9bea-9071d8b41cb8", "address": "fa:16:3e:2e:cc:1b", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8824a231-01", "ovs_interfaceid": "8824a231-01df-41e5-9bea-9071d8b41cb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.347723] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ced84d-be2c-4433-8dfb-aa4d9f774207 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.378823] env[62521]: DEBUG nova.network.neutron [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance_info_cache with network_info: [{"id": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "address": "fa:16:3e:d7:52:1f", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e00f1c5-5c", "ovs_interfaceid": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.689208] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.689734] env[62521]: DEBUG nova.compute.manager [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 865.693694] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.774s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.693944] env[62521]: DEBUG nova.objects.instance [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lazy-loading 'resources' on Instance uuid 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.797012] env[62521]: DEBUG nova.compute.manager [req-0d6f8a48-6f82-4ac5-805d-3bc44f6c4e05 req-259a6b9d-d93d-4907-b16d-9119b9fd0500 service nova] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Received event network-vif-plugged-1643da59-fa10-4645-994b-0c557c3a6e10 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.797445] env[62521]: DEBUG oslo_concurrency.lockutils [req-0d6f8a48-6f82-4ac5-805d-3bc44f6c4e05 req-259a6b9d-d93d-4907-b16d-9119b9fd0500 service nova] Acquiring lock "3805d897-fcb5-43ed-b502-7fad71124e74-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.797726] env[62521]: DEBUG oslo_concurrency.lockutils [req-0d6f8a48-6f82-4ac5-805d-3bc44f6c4e05 req-259a6b9d-d93d-4907-b16d-9119b9fd0500 service nova] Lock "3805d897-fcb5-43ed-b502-7fad71124e74-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.798009] env[62521]: DEBUG oslo_concurrency.lockutils [req-0d6f8a48-6f82-4ac5-805d-3bc44f6c4e05 req-259a6b9d-d93d-4907-b16d-9119b9fd0500 service nova] Lock "3805d897-fcb5-43ed-b502-7fad71124e74-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.798727] env[62521]: DEBUG nova.compute.manager [req-0d6f8a48-6f82-4ac5-805d-3bc44f6c4e05 req-259a6b9d-d93d-4907-b16d-9119b9fd0500 service nova] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] No waiting events found dispatching network-vif-plugged-1643da59-fa10-4645-994b-0c557c3a6e10 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.799150] env[62521]: WARNING nova.compute.manager [req-0d6f8a48-6f82-4ac5-805d-3bc44f6c4e05 req-259a6b9d-d93d-4907-b16d-9119b9fd0500 service nova] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Received unexpected event network-vif-plugged-1643da59-fa10-4645-994b-0c557c3a6e10 for instance with vm_state building and task_state spawning. [ 865.843164] env[62521]: INFO nova.compute.manager [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Took 0.57 seconds to detach 1 volumes for instance. [ 865.848045] env[62521]: DEBUG oslo_concurrency.lockutils [req-c4f12dc2-fc3b-43a7-9ca9-6b65b9b6b1db req-926f0e60-fbd1-4c9d-b35a-5ba5a8de1526 service nova] Releasing lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.850372] env[62521]: DEBUG nova.compute.manager [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Deleting volume: 683a2b16-e997-4b24-90c9-559ca2510a58 {{(pid=62521) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 865.882280] env[62521]: DEBUG oslo_concurrency.lockutils [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.909076] env[62521]: DEBUG nova.network.neutron [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Successfully updated port: 1643da59-fa10-4645-994b-0c557c3a6e10 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 866.196364] env[62521]: DEBUG nova.compute.utils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.198463] env[62521]: DEBUG nova.compute.manager [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 866.198463] env[62521]: DEBUG nova.network.neutron [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 866.250667] env[62521]: DEBUG nova.policy [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f044b14654f45a9b40143285f4b1672', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a592d0f3c5834858bda6ca89a33662a9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 866.298427] env[62521]: DEBUG nova.compute.manager [req-dfff90ba-c693-40d6-a2a5-3c9a2b892b57 req-78374c64-7c54-42c0-85f3-dfa66f1e0f01 service nova] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Received event network-vif-deleted-53471722-7f78-46af-bc13-02c61f597490 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.414221] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.414379] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.414529] env[62521]: DEBUG nova.network.neutron [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 866.416680] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428a3d0a-75ff-49f8-ba68-8a6c3e200962 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.425945] env[62521]: DEBUG oslo_concurrency.lockutils [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.429417] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 866.429566] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282105', 'volume_id': '6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'name': 'volume-6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '538ed18e-920e-4ed4-a9e1-a6bfbae7f282', 'attached_at': '', 'detached_at': '', 'volume_id': '6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'serial': '6e8eee93-7abc-4c71-8667-16b6b4dbd033'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 866.430621] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6002dc-f0b3-4bd9-8f71-330d11d8d035 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.452199] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ec52d3-2fe7-4457-975c-146ecc072230 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.467288] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cefc89-994f-4475-89fb-b8514a8fe1b0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.472754] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance 'd5f6492d-ca1b-4e74-b792-b04d55c33660' progress to 83 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 866.499996] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] volume-6e8eee93-7abc-4c71-8667-16b6b4dbd033/volume-6e8eee93-7abc-4c71-8667-16b6b4dbd033.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.502950] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70d6c3d0-68c0-485d-9179-1bd5a2c14ddd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.523494] env[62521]: DEBUG oslo_vmware.api [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 866.523494] env[62521]: value = "task-1318631" [ 866.523494] env[62521]: _type = "Task" [ 866.523494] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.534726] env[62521]: DEBUG oslo_vmware.api [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318631, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.552279] env[62521]: DEBUG nova.network.neutron [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Successfully created port: 9a829e1f-fed8-4e5c-a3c1-481e993b2551 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 866.601873] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf38efd-65ba-4cb8-9f0b-35b56c252cd6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.609441] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392013e2-ad7d-4d8c-923c-e6dc3b794a88 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.639221] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a356c797-f699-4cd2-a4e9-64314e5cc2f7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.646862] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e964de-d85e-40b0-bf64-ce6d15eded65 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.659866] env[62521]: DEBUG nova.compute.provider_tree [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.701299] env[62521]: DEBUG nova.compute.manager [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 866.956361] env[62521]: DEBUG nova.network.neutron [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.980811] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.981178] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8144065e-c47e-4350-9726-73040ed2966f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.988780] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 866.988780] env[62521]: value = "task-1318632" [ 866.988780] env[62521]: _type = "Task" [ 866.988780] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.996314] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318632, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.059775] env[62521]: DEBUG oslo_vmware.api [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318631, 'name': ReconfigVM_Task, 'duration_secs': 0.363397} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.059775] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Reconfigured VM instance instance-00000037 to attach disk [datastore2] volume-6e8eee93-7abc-4c71-8667-16b6b4dbd033/volume-6e8eee93-7abc-4c71-8667-16b6b4dbd033.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.064454] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1079984-b4e0-44dc-9eab-dde9a470eaa0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.080172] env[62521]: DEBUG oslo_vmware.api [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 867.080172] env[62521]: value = "task-1318633" [ 867.080172] env[62521]: _type = "Task" [ 867.080172] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.089923] env[62521]: DEBUG oslo_vmware.api [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318633, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.163358] env[62521]: DEBUG nova.scheduler.client.report [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.227472] env[62521]: DEBUG nova.network.neutron [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance_info_cache with network_info: [{"id": "1643da59-fa10-4645-994b-0c557c3a6e10", "address": "fa:16:3e:11:0f:18", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1643da59-fa", "ovs_interfaceid": "1643da59-fa10-4645-994b-0c557c3a6e10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.330165] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 867.330531] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 867.500210] env[62521]: DEBUG oslo_vmware.api [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318632, 'name': PowerOnVM_Task, 'duration_secs': 0.430253} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.500490] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.500681] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-663d7e5d-6cc5-4cc0-aa2b-48b0d883fb13 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance 'd5f6492d-ca1b-4e74-b792-b04d55c33660' progress to 100 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 867.590872] env[62521]: DEBUG oslo_vmware.api [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318633, 'name': ReconfigVM_Task, 'duration_secs': 0.157373} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.590872] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282105', 'volume_id': '6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'name': 'volume-6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '538ed18e-920e-4ed4-a9e1-a6bfbae7f282', 'attached_at': '', 'detached_at': '', 'volume_id': '6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'serial': '6e8eee93-7abc-4c71-8667-16b6b4dbd033'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 867.668145] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.974s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.670648] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.803s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.670892] env[62521]: DEBUG nova.objects.instance [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lazy-loading 'resources' on Instance uuid 583b1a1b-1817-45fc-aa4f-fc9de4b4a243 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.690031] env[62521]: INFO nova.scheduler.client.report [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted allocations for instance 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6 [ 867.711696] env[62521]: DEBUG nova.compute.manager [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 867.730479] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.730800] env[62521]: DEBUG nova.compute.manager [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Instance network_info: |[{"id": "1643da59-fa10-4645-994b-0c557c3a6e10", "address": "fa:16:3e:11:0f:18", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1643da59-fa", "ovs_interfaceid": "1643da59-fa10-4645-994b-0c557c3a6e10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 867.733552] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:0f:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1643da59-fa10-4645-994b-0c557c3a6e10', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.741516] env[62521]: DEBUG oslo.service.loopingcall [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.742226] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.742473] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a4dff0c-c27c-4b21-baf5-6a08d5248e77 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.758906] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.759154] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.759323] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.759515] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.759666] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.759818] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.760076] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.760271] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.760451] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.760618] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.760795] env[62521]: DEBUG nova.virt.hardware [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.761846] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6d8a05-ff58-48e7-8ca5-e16e2e02fde7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.770829] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528c7691-8ec2-47dd-87e9-92f8a8aedec5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.774365] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.774365] env[62521]: value = "task-1318634" [ 867.774365] env[62521]: _type = "Task" [ 867.774365] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.790779] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318634, 'name': CreateVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.824205] env[62521]: DEBUG nova.compute.manager [req-4d1abd98-2dcd-4442-a6c1-3a371c4dfea4 req-1dd861fc-e294-45ad-9afe-e8e920db8322 service nova] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Received event network-changed-1643da59-fa10-4645-994b-0c557c3a6e10 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.824368] env[62521]: DEBUG nova.compute.manager [req-4d1abd98-2dcd-4442-a6c1-3a371c4dfea4 req-1dd861fc-e294-45ad-9afe-e8e920db8322 service nova] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Refreshing instance network info cache due to event network-changed-1643da59-fa10-4645-994b-0c557c3a6e10. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 867.824629] env[62521]: DEBUG oslo_concurrency.lockutils [req-4d1abd98-2dcd-4442-a6c1-3a371c4dfea4 req-1dd861fc-e294-45ad-9afe-e8e920db8322 service nova] Acquiring lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.824700] env[62521]: DEBUG oslo_concurrency.lockutils [req-4d1abd98-2dcd-4442-a6c1-3a371c4dfea4 req-1dd861fc-e294-45ad-9afe-e8e920db8322 service nova] Acquired lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.824875] env[62521]: DEBUG nova.network.neutron [req-4d1abd98-2dcd-4442-a6c1-3a371c4dfea4 req-1dd861fc-e294-45ad-9afe-e8e920db8322 service nova] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Refreshing network info cache for port 1643da59-fa10-4645-994b-0c557c3a6e10 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.838804] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 867.840089] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 867.840089] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Rebuilding the list of instances to heal {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 868.197332] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a9c9e0f4-185b-488c-934f-613fcfacb297 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.379s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.285585] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318634, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.292268] env[62521]: DEBUG nova.network.neutron [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Successfully updated port: 9a829e1f-fed8-4e5c-a3c1-481e993b2551 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.344649] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 868.344733] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 868.402990] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.403154] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquired lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.403304] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Forcefully refreshing network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 868.403464] env[62521]: DEBUG nova.objects.instance [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lazy-loading 'info_cache' on Instance uuid 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.442893] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0777083-3b57-4c1a-8c11-6a2cbb2c0aa4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.450261] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b7d776-15e8-4d15-8dd2-4b262424b2d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.482494] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbd8815-a746-433b-9559-ff11bbff75ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.490245] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685953d4-141a-4ee8-b119-e80ca4fb4cd6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.504740] env[62521]: DEBUG nova.compute.provider_tree [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.627197] env[62521]: DEBUG nova.objects.instance [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'flavor' on Instance uuid 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.629900] env[62521]: DEBUG nova.network.neutron [req-4d1abd98-2dcd-4442-a6c1-3a371c4dfea4 req-1dd861fc-e294-45ad-9afe-e8e920db8322 service nova] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updated VIF entry in instance network info cache for port 1643da59-fa10-4645-994b-0c557c3a6e10. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.630270] env[62521]: DEBUG nova.network.neutron [req-4d1abd98-2dcd-4442-a6c1-3a371c4dfea4 req-1dd861fc-e294-45ad-9afe-e8e920db8322 service nova] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance_info_cache with network_info: [{"id": "1643da59-fa10-4645-994b-0c557c3a6e10", "address": "fa:16:3e:11:0f:18", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1643da59-fa", "ovs_interfaceid": "1643da59-fa10-4645-994b-0c557c3a6e10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.785804] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318634, 'name': CreateVM_Task, 'duration_secs': 0.74199} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.786013] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 868.786745] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.786924] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.787287] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.787550] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52e27b9e-9ca2-4e0b-a6f9-0a4156de98d3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.791960] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 868.791960] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525fe34d-3067-68bf-91c5-a0b3ae064521" [ 868.791960] env[62521]: _type = "Task" [ 868.791960] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.795102] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "refresh_cache-acfd78aa-a51e-4a1c-b01a-15854fe0635c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.795261] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "refresh_cache-acfd78aa-a51e-4a1c-b01a-15854fe0635c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.795440] env[62521]: DEBUG nova.network.neutron [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.801880] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525fe34d-3067-68bf-91c5-a0b3ae064521, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.009977] env[62521]: DEBUG nova.scheduler.client.report [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.131994] env[62521]: DEBUG oslo_concurrency.lockutils [None req-adb6f433-32e7-44ac-885d-796c4854afb7 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.332s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.133459] env[62521]: DEBUG oslo_concurrency.lockutils [req-4d1abd98-2dcd-4442-a6c1-3a371c4dfea4 req-1dd861fc-e294-45ad-9afe-e8e920db8322 service nova] Releasing lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.304102] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525fe34d-3067-68bf-91c5-a0b3ae064521, 'name': SearchDatastore_Task, 'duration_secs': 0.011727} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.304738] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.304984] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 869.305250] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.305405] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.305594] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 869.305857] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27eb6cb2-4cd0-46f4-acd2-0d435d0eca4f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.313521] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 869.313605] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 869.314329] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b26bf73b-9b1e-4c8c-a555-44785fd21275 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.323296] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 869.323296] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523f6e2a-abbd-d561-86a6-554cdc58a174" [ 869.323296] env[62521]: _type = "Task" [ 869.323296] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.331591] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523f6e2a-abbd-d561-86a6-554cdc58a174, 'name': SearchDatastore_Task, 'duration_secs': 0.00909} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.334316] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0171a7d-57f0-48a8-9999-1301b4eb2028 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.339300] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 869.339300] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526d65e5-d87d-b026-9541-0570f0e91e97" [ 869.339300] env[62521]: _type = "Task" [ 869.339300] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.340141] env[62521]: DEBUG nova.network.neutron [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.350436] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526d65e5-d87d-b026-9541-0570f0e91e97, 'name': SearchDatastore_Task, 'duration_secs': 0.00873} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.350772] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.350944] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 3805d897-fcb5-43ed-b502-7fad71124e74/3805d897-fcb5-43ed-b502-7fad71124e74.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 869.351274] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c658eaa-52e7-4d71-8ad3-4978668f9db5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.357951] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 869.357951] env[62521]: value = "task-1318635" [ 869.357951] env[62521]: _type = "Task" [ 869.357951] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.366168] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318635, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.515607] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.845s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.518230] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.411s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.521667] env[62521]: INFO nova.compute.claims [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.557195] env[62521]: INFO nova.scheduler.client.report [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Deleted allocations for instance 583b1a1b-1817-45fc-aa4f-fc9de4b4a243 [ 869.587507] env[62521]: DEBUG nova.network.neutron [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Updating instance_info_cache with network_info: [{"id": "9a829e1f-fed8-4e5c-a3c1-481e993b2551", "address": "fa:16:3e:77:c0:f5", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a829e1f-fe", "ovs_interfaceid": "9a829e1f-fed8-4e5c-a3c1-481e993b2551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.597937] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.598357] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.598578] env[62521]: DEBUG nova.compute.manager [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.600638] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-becefe81-881f-4c08-9a3d-590f10a7d6dd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.609968] env[62521]: DEBUG nova.compute.manager [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62521) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 869.610642] env[62521]: DEBUG nova.objects.instance [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'flavor' on Instance uuid 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.852996] env[62521]: DEBUG nova.compute.manager [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Received event network-vif-plugged-9a829e1f-fed8-4e5c-a3c1-481e993b2551 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.853278] env[62521]: DEBUG oslo_concurrency.lockutils [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] Acquiring lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.853843] env[62521]: DEBUG oslo_concurrency.lockutils [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] Lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.854041] env[62521]: DEBUG oslo_concurrency.lockutils [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] Lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.854231] env[62521]: DEBUG nova.compute.manager [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] No waiting events found dispatching network-vif-plugged-9a829e1f-fed8-4e5c-a3c1-481e993b2551 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.854408] env[62521]: WARNING nova.compute.manager [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Received unexpected event network-vif-plugged-9a829e1f-fed8-4e5c-a3c1-481e993b2551 for instance with vm_state building and task_state spawning. [ 869.854578] env[62521]: DEBUG nova.compute.manager [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Received event network-changed-9a829e1f-fed8-4e5c-a3c1-481e993b2551 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.854760] env[62521]: DEBUG nova.compute.manager [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Refreshing instance network info cache due to event network-changed-9a829e1f-fed8-4e5c-a3c1-481e993b2551. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 869.855271] env[62521]: DEBUG oslo_concurrency.lockutils [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] Acquiring lock "refresh_cache-acfd78aa-a51e-4a1c-b01a-15854fe0635c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.870556] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318635, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491336} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.870556] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 3805d897-fcb5-43ed-b502-7fad71124e74/3805d897-fcb5-43ed-b502-7fad71124e74.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 869.870556] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 869.870556] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec1455e2-fd6b-4a48-a688-266836ffbe84 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.875591] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 869.875591] env[62521]: value = "task-1318636" [ 869.875591] env[62521]: _type = "Task" [ 869.875591] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.883580] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318636, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.066831] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b1e177c-d16b-4585-b6bb-b4745f63437b tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "583b1a1b-1817-45fc-aa4f-fc9de4b4a243" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.680s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.090055] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "refresh_cache-acfd78aa-a51e-4a1c-b01a-15854fe0635c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.090373] env[62521]: DEBUG nova.compute.manager [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Instance network_info: |[{"id": "9a829e1f-fed8-4e5c-a3c1-481e993b2551", "address": "fa:16:3e:77:c0:f5", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a829e1f-fe", "ovs_interfaceid": "9a829e1f-fed8-4e5c-a3c1-481e993b2551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 870.090967] env[62521]: DEBUG oslo_concurrency.lockutils [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] Acquired lock "refresh_cache-acfd78aa-a51e-4a1c-b01a-15854fe0635c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.090967] env[62521]: DEBUG nova.network.neutron [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Refreshing network info cache for port 9a829e1f-fed8-4e5c-a3c1-481e993b2551 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.092293] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:c0:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd646f9d5-d2ad-4c22-bea5-85a965334de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a829e1f-fed8-4e5c-a3c1-481e993b2551', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.099924] env[62521]: DEBUG oslo.service.loopingcall [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.100365] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.101199] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc65404e-f4b3-4572-8345-98bd56b0a32e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.120804] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 870.121857] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eef3c228-c435-4fc6-b21b-630bda93d04c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.125839] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.125839] env[62521]: value = "task-1318637" [ 870.125839] env[62521]: _type = "Task" [ 870.125839] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.129925] env[62521]: DEBUG oslo_vmware.api [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 870.129925] env[62521]: value = "task-1318638" [ 870.129925] env[62521]: _type = "Task" [ 870.129925] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.135967] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318637, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.140461] env[62521]: DEBUG oslo_vmware.api [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.181608] env[62521]: DEBUG oslo_concurrency.lockutils [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "d5f6492d-ca1b-4e74-b792-b04d55c33660" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.182070] env[62521]: DEBUG oslo_concurrency.lockutils [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.182210] env[62521]: DEBUG nova.compute.manager [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Going to confirm migration 1 {{(pid=62521) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 870.193363] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updating instance_info_cache with network_info: [{"id": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "address": "fa:16:3e:41:9c:b1", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06aa91ed-3b", "ovs_interfaceid": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.394700] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318636, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073883} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.395055] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.395807] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27dddff9-c6f2-4986-90cf-4d5c0c2380b0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.423229] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 3805d897-fcb5-43ed-b502-7fad71124e74/3805d897-fcb5-43ed-b502-7fad71124e74.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.423583] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e70aa6d-21d8-4507-91de-7ba241493df9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.448686] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 870.448686] env[62521]: value = "task-1318639" [ 870.448686] env[62521]: _type = "Task" [ 870.448686] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.457183] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318639, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.640969] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318637, 'name': CreateVM_Task, 'duration_secs': 0.481238} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.645869] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.646236] env[62521]: DEBUG oslo_vmware.api [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318638, 'name': PowerOffVM_Task, 'duration_secs': 0.323947} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.647063] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.647282] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.647557] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.648053] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.648053] env[62521]: DEBUG nova.compute.manager [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.648260] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fb087ac-313c-4377-836d-0578e7e542d1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.650253] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e982789b-7a5e-4d29-81ae-af2ebd586a2f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.655393] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 870.655393] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5274274a-e066-0a7f-e14c-8b4404a1c342" [ 870.655393] env[62521]: _type = "Task" [ 870.655393] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.669930] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5274274a-e066-0a7f-e14c-8b4404a1c342, 'name': SearchDatastore_Task, 'duration_secs': 0.009172} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.669930] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.669930] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.670151] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.670304] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.670558] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.670745] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ddcf9708-e300-4a86-b5a5-c326fe5b1211 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.679267] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.680026] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 870.680200] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22034fd6-068d-4ba9-8741-5a578ec8bd7a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.687936] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 870.687936] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522900e0-ff12-3a3b-7bf0-09cb0c1edbdf" [ 870.687936] env[62521]: _type = "Task" [ 870.687936] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.698189] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Releasing lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.698396] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updated the network info_cache for instance {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 870.698593] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.699396] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.699625] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.699974] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.700320] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.704441] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.704592] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 870.705151] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522900e0-ff12-3a3b-7bf0-09cb0c1edbdf, 'name': SearchDatastore_Task, 'duration_secs': 0.008578} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.707504] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.709307] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d26b46f-7f07-49fd-8295-19f97b2757f6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.714528] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 870.714528] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5264aa47-7117-c6eb-b481-c486f4f64b1b" [ 870.714528] env[62521]: _type = "Task" [ 870.714528] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.724435] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5264aa47-7117-c6eb-b481-c486f4f64b1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.764749] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f87561-a393-4777-971f-696aaa142edb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.772138] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63b6bfd-727c-4b26-ad20-c31f0afa5b70 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.803633] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472f83f6-0226-4880-9d1f-b0df4e85ece6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.810873] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30efc37e-d442-459c-8de8-0acf1ac31696 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.823892] env[62521]: DEBUG nova.compute.provider_tree [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.959255] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318639, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.042517] env[62521]: DEBUG oslo_concurrency.lockutils [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.042751] env[62521]: DEBUG oslo_concurrency.lockutils [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.089667] env[62521]: DEBUG oslo_concurrency.lockutils [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.089915] env[62521]: DEBUG oslo_concurrency.lockutils [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.090044] env[62521]: DEBUG nova.network.neutron [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.090251] env[62521]: DEBUG nova.objects.instance [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lazy-loading 'info_cache' on Instance uuid d5f6492d-ca1b-4e74-b792-b04d55c33660 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.151571] env[62521]: DEBUG nova.network.neutron [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Updated VIF entry in instance network info cache for port 9a829e1f-fed8-4e5c-a3c1-481e993b2551. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.153294] env[62521]: DEBUG nova.network.neutron [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Updating instance_info_cache with network_info: [{"id": "9a829e1f-fed8-4e5c-a3c1-481e993b2551", "address": "fa:16:3e:77:c0:f5", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a829e1f-fe", "ovs_interfaceid": "9a829e1f-fed8-4e5c-a3c1-481e993b2551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.167800] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6539f23-4863-4c3e-87b8-4a6c484ec8c1 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.569s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.211256] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.225079] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5264aa47-7117-c6eb-b481-c486f4f64b1b, 'name': SearchDatastore_Task, 'duration_secs': 0.008437} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.225381] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.225645] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] acfd78aa-a51e-4a1c-b01a-15854fe0635c/acfd78aa-a51e-4a1c-b01a-15854fe0635c.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 871.226043] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27aeb3ee-7307-4695-ac6f-0abdced30544 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.232488] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 871.232488] env[62521]: value = "task-1318640" [ 871.232488] env[62521]: _type = "Task" [ 871.232488] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.240179] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318640, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.327546] env[62521]: DEBUG nova.scheduler.client.report [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.461506] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318639, 'name': ReconfigVM_Task, 'duration_secs': 0.580135} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.461914] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 3805d897-fcb5-43ed-b502-7fad71124e74/3805d897-fcb5-43ed-b502-7fad71124e74.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.462670] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4ffd89d-e9d7-419b-9796-9dfeeccb1995 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.469551] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 871.469551] env[62521]: value = "task-1318641" [ 871.469551] env[62521]: _type = "Task" [ 871.469551] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.479178] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318641, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.545802] env[62521]: DEBUG nova.compute.utils [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 871.655037] env[62521]: DEBUG oslo_concurrency.lockutils [req-73849c67-c15a-4470-bcd6-42036ab21379 req-36003047-940a-47ce-b525-f1a41796117a service nova] Releasing lock "refresh_cache-acfd78aa-a51e-4a1c-b01a-15854fe0635c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.743301] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318640, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.833592] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.834281] env[62521]: DEBUG nova.compute.manager [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.837258] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.585s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.837503] env[62521]: DEBUG nova.objects.instance [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lazy-loading 'resources' on Instance uuid 92629ebf-eb50-4862-abe2-01e210e302ad {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.980184] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318641, 'name': Rename_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.048617] env[62521]: DEBUG oslo_concurrency.lockutils [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.243320] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318640, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596262} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.243912] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] acfd78aa-a51e-4a1c-b01a-15854fe0635c/acfd78aa-a51e-4a1c-b01a-15854fe0635c.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 872.243912] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.244162] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a632b2c-b619-40ac-9bff-db513a443b86 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.252551] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 872.252551] env[62521]: value = "task-1318642" [ 872.252551] env[62521]: _type = "Task" [ 872.252551] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.260455] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318642, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.340093] env[62521]: DEBUG nova.compute.utils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.343963] env[62521]: DEBUG nova.compute.manager [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.346017] env[62521]: DEBUG nova.network.neutron [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 872.347564] env[62521]: DEBUG nova.objects.instance [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'flavor' on Instance uuid 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.403017] env[62521]: DEBUG nova.network.neutron [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance_info_cache with network_info: [{"id": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "address": "fa:16:3e:d7:52:1f", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e00f1c5-5c", "ovs_interfaceid": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.410434] env[62521]: DEBUG nova.policy [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76d85009c2c1416b9ea049003e6d96a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c533bd4917b4466e9c5c76a9caf8e807', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.484249] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318641, 'name': Rename_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.642680] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ae8694-4967-4833-bd6e-534361bea18d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.651778] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9ae3ad-7ead-40c8-a167-8304e853cc91 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.689480] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a454208c-ff1a-481e-a3b0-10c3d97e5425 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.698193] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b4ab93-374c-406b-a9dc-68460255f2dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.718834] env[62521]: DEBUG nova.compute.provider_tree [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 872.767106] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318642, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.339338} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.767398] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 872.768200] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8ee625-0ada-418c-bc7c-90196e1f379a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.790213] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] acfd78aa-a51e-4a1c-b01a-15854fe0635c/acfd78aa-a51e-4a1c-b01a-15854fe0635c.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.790474] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0735b1e6-409a-4d8b-a8e3-78721bff5d2d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.810073] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 872.810073] env[62521]: value = "task-1318643" [ 872.810073] env[62521]: _type = "Task" [ 872.810073] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.819093] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318643, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.844836] env[62521]: DEBUG nova.compute.manager [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.851116] env[62521]: DEBUG nova.network.neutron [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Successfully created port: 862343de-1cfa-4880-832a-bc8881cd0c30 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.852822] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.852984] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquired lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.853166] env[62521]: DEBUG nova.network.neutron [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 872.853339] env[62521]: DEBUG nova.objects.instance [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'info_cache' on Instance uuid 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.909678] env[62521]: DEBUG oslo_concurrency.lockutils [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.910013] env[62521]: DEBUG nova.objects.instance [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lazy-loading 'migration_context' on Instance uuid d5f6492d-ca1b-4e74-b792-b04d55c33660 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.980467] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318641, 'name': Rename_Task, 'duration_secs': 1.2441} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.980775] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 872.981042] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a6660869-a7cb-4239-bcf3-6f3d73cf94a4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.987515] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 872.987515] env[62521]: value = "task-1318644" [ 872.987515] env[62521]: _type = "Task" [ 872.987515] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.995289] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318644, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.075681] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "6c4c36b7-b242-4578-9c23-d2529f308cb1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.076037] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "6c4c36b7-b242-4578-9c23-d2529f308cb1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.161461] env[62521]: DEBUG oslo_concurrency.lockutils [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.161575] env[62521]: DEBUG oslo_concurrency.lockutils [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.161827] env[62521]: INFO nova.compute.manager [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Attaching volume 5e701085-e961-46c7-8528-8e06b1899590 to /dev/sdb [ 873.197508] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458d9abe-d9cf-40f1-b56c-9614d2ce4640 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.204278] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ba3049-5d54-431b-8751-de388f5ae4a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.220497] env[62521]: DEBUG nova.virt.block_device [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Updating existing volume attachment record: dde5ccb5-ffb1-40f2-a40a-0c7c8a5856e7 {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 873.241923] env[62521]: ERROR nova.scheduler.client.report [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] [req-b11c35fb-83fc-431b-bd39-7cc5757d1895] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b11c35fb-83fc-431b-bd39-7cc5757d1895"}]} [ 873.258441] env[62521]: DEBUG nova.scheduler.client.report [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 873.272848] env[62521]: DEBUG nova.scheduler.client.report [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 873.273170] env[62521]: DEBUG nova.compute.provider_tree [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 873.285336] env[62521]: DEBUG nova.scheduler.client.report [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 873.303625] env[62521]: DEBUG nova.scheduler.client.report [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 873.320477] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318643, 'name': ReconfigVM_Task, 'duration_secs': 0.31585} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.320477] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Reconfigured VM instance instance-00000046 to attach disk [datastore2] acfd78aa-a51e-4a1c-b01a-15854fe0635c/acfd78aa-a51e-4a1c-b01a-15854fe0635c.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.320929] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-effd3134-fd4f-4da2-9479-746fa7e41ce9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.329294] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 873.329294] env[62521]: value = "task-1318645" [ 873.329294] env[62521]: _type = "Task" [ 873.329294] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.338352] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318645, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.356395] env[62521]: DEBUG nova.objects.base [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Object Instance<538ed18e-920e-4ed4-a9e1-a6bfbae7f282> lazy-loaded attributes: flavor,info_cache {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 873.412549] env[62521]: DEBUG nova.objects.base [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 873.413997] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11c6183-7ecd-40bc-88c4-ad3b6c9f95e0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.437024] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb418158-49ec-4dc9-b6e4-9092f949e8bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.443456] env[62521]: DEBUG oslo_vmware.api [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 873.443456] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52670112-99de-8039-747f-08d046da21bc" [ 873.443456] env[62521]: _type = "Task" [ 873.443456] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.452087] env[62521]: DEBUG oslo_vmware.api [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52670112-99de-8039-747f-08d046da21bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.497362] env[62521]: DEBUG oslo_vmware.api [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318644, 'name': PowerOnVM_Task, 'duration_secs': 0.439484} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.499691] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.499908] env[62521]: INFO nova.compute.manager [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Took 8.21 seconds to spawn the instance on the hypervisor. [ 873.500114] env[62521]: DEBUG nova.compute.manager [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.501097] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00139a8c-dde2-4b83-867b-600ef8d46364 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.580273] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b28faa23-ed42-40e4-bc4b-1ff56756ef8c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.583855] env[62521]: DEBUG nova.compute.manager [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.591668] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a59d0cc-94cb-4034-98eb-f4266a1f1e85 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.625044] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5cc3561-c186-486b-8d66-7f224ee00aa7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.633958] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6270cf32-711f-425f-a559-b87e03734f60 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.647286] env[62521]: DEBUG nova.compute.provider_tree [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 873.840026] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318645, 'name': Rename_Task, 'duration_secs': 0.138098} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.840183] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.840436] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ade9adfe-d556-4694-b3d1-6668c5269d23 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.847040] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 873.847040] env[62521]: value = "task-1318649" [ 873.847040] env[62521]: _type = "Task" [ 873.847040] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.855142] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318649, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.856338] env[62521]: DEBUG nova.compute.manager [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.881796] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.882102] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.882285] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.882499] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.882663] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.882865] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.883218] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.883403] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.883629] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.883961] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.884056] env[62521]: DEBUG nova.virt.hardware [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.884970] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e429d94-3135-4d18-a2a4-6435c39cb06f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.895016] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b56aa62-f8dd-4be6-8ef7-9da8151f8d7f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.960837] env[62521]: DEBUG oslo_vmware.api [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52670112-99de-8039-747f-08d046da21bc, 'name': SearchDatastore_Task, 'duration_secs': 0.02159} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.961303] env[62521]: DEBUG oslo_concurrency.lockutils [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.022531] env[62521]: INFO nova.compute.manager [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Took 19.57 seconds to build instance. [ 874.105930] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.182586] env[62521]: DEBUG nova.scheduler.client.report [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 87 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 874.183595] env[62521]: DEBUG nova.compute.provider_tree [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 87 to 88 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 874.183595] env[62521]: DEBUG nova.compute.provider_tree [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 874.211334] env[62521]: DEBUG nova.network.neutron [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updating instance_info_cache with network_info: [{"id": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "address": "fa:16:3e:41:9c:b1", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06aa91ed-3b", "ovs_interfaceid": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.357804] env[62521]: DEBUG oslo_vmware.api [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318649, 'name': PowerOnVM_Task, 'duration_secs': 0.471922} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.357804] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.358032] env[62521]: INFO nova.compute.manager [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Took 6.65 seconds to spawn the instance on the hypervisor. [ 874.360404] env[62521]: DEBUG nova.compute.manager [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.360404] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85643e21-8d13-4cc2-8179-0f2aa20dd93f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.468844] env[62521]: DEBUG nova.compute.manager [req-97468314-16ee-4f82-b7d7-5c50dbf6c14a req-b7ad43d7-5f26-479a-8875-735dfcc59ba6 service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Received event network-vif-plugged-862343de-1cfa-4880-832a-bc8881cd0c30 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.468844] env[62521]: DEBUG oslo_concurrency.lockutils [req-97468314-16ee-4f82-b7d7-5c50dbf6c14a req-b7ad43d7-5f26-479a-8875-735dfcc59ba6 service nova] Acquiring lock "d574f077-b221-4f91-8b54-0915421cb36f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.468844] env[62521]: DEBUG oslo_concurrency.lockutils [req-97468314-16ee-4f82-b7d7-5c50dbf6c14a req-b7ad43d7-5f26-479a-8875-735dfcc59ba6 service nova] Lock "d574f077-b221-4f91-8b54-0915421cb36f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.468844] env[62521]: DEBUG oslo_concurrency.lockutils [req-97468314-16ee-4f82-b7d7-5c50dbf6c14a req-b7ad43d7-5f26-479a-8875-735dfcc59ba6 service nova] Lock "d574f077-b221-4f91-8b54-0915421cb36f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.468844] env[62521]: DEBUG nova.compute.manager [req-97468314-16ee-4f82-b7d7-5c50dbf6c14a req-b7ad43d7-5f26-479a-8875-735dfcc59ba6 service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] No waiting events found dispatching network-vif-plugged-862343de-1cfa-4880-832a-bc8881cd0c30 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.468844] env[62521]: WARNING nova.compute.manager [req-97468314-16ee-4f82-b7d7-5c50dbf6c14a req-b7ad43d7-5f26-479a-8875-735dfcc59ba6 service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Received unexpected event network-vif-plugged-862343de-1cfa-4880-832a-bc8881cd0c30 for instance with vm_state building and task_state spawning. [ 874.523273] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e0cf6ea0-6f30-46bc-bb11-d1ba7e9505da tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "3805d897-fcb5-43ed-b502-7fad71124e74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.082s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.621168] env[62521]: DEBUG nova.network.neutron [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Successfully updated port: 862343de-1cfa-4880-832a-bc8881cd0c30 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.688661] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.851s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.691611] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.455s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.693142] env[62521]: INFO nova.compute.claims [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.713553] env[62521]: INFO nova.scheduler.client.report [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Deleted allocations for instance 92629ebf-eb50-4862-abe2-01e210e302ad [ 874.715029] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Releasing lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.878509] env[62521]: INFO nova.compute.manager [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Took 19.00 seconds to build instance. [ 875.124872] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "refresh_cache-d574f077-b221-4f91-8b54-0915421cb36f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.124872] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "refresh_cache-d574f077-b221-4f91-8b54-0915421cb36f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.124872] env[62521]: DEBUG nova.network.neutron [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.220313] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.222782] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91e2d415-b84d-44d0-bc83-16a303b56d51 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.230533] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4c61b970-0920-4064-a278-622ef3060141 tempest-InstanceActionsTestJSON-984996532 tempest-InstanceActionsTestJSON-984996532-project-member] Lock "92629ebf-eb50-4862-abe2-01e210e302ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.400s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.237083] env[62521]: DEBUG oslo_vmware.api [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 875.237083] env[62521]: value = "task-1318650" [ 875.237083] env[62521]: _type = "Task" [ 875.237083] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.246671] env[62521]: DEBUG oslo_vmware.api [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.380418] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6176f3a0-860a-45ce-816f-23782c5b152c tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.510s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.663318] env[62521]: DEBUG nova.network.neutron [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.757184] env[62521]: DEBUG oslo_vmware.api [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318650, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.784540] env[62521]: DEBUG oslo_concurrency.lockutils [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.784790] env[62521]: DEBUG oslo_concurrency.lockutils [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.784995] env[62521]: DEBUG oslo_concurrency.lockutils [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.785204] env[62521]: DEBUG oslo_concurrency.lockutils [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.785378] env[62521]: DEBUG oslo_concurrency.lockutils [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.791378] env[62521]: INFO nova.compute.manager [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Terminating instance [ 875.794023] env[62521]: DEBUG nova.compute.manager [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 875.794332] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 875.795566] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2108a07f-9b4b-4779-be71-2feeba76095f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.805891] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 875.809435] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20883052-130d-49e1-b5aa-0b4c61490eb4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.818338] env[62521]: DEBUG oslo_vmware.api [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 875.818338] env[62521]: value = "task-1318652" [ 875.818338] env[62521]: _type = "Task" [ 875.818338] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.829950] env[62521]: DEBUG oslo_vmware.api [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318652, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.951450] env[62521]: DEBUG nova.network.neutron [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Updating instance_info_cache with network_info: [{"id": "862343de-1cfa-4880-832a-bc8881cd0c30", "address": "fa:16:3e:d5:4b:bc", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap862343de-1c", "ovs_interfaceid": "862343de-1cfa-4880-832a-bc8881cd0c30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.017724] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23549ff2-206a-4c13-b115-b25c58f62953 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.026139] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923a79ce-c7da-43f3-ba73-2e183380bbf0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.057968] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8300c9-af8b-428e-b736-4b1d3e517859 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.066265] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d684b8e4-f8f8-4afb-a0ee-6aa5c8dafc53 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.082020] env[62521]: DEBUG nova.compute.provider_tree [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.208054] env[62521]: DEBUG nova.compute.manager [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Stashing vm_state: active {{(pid=62521) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 876.248967] env[62521]: DEBUG oslo_vmware.api [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318650, 'name': PowerOnVM_Task, 'duration_secs': 0.712687} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.249538] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 876.250061] env[62521]: DEBUG nova.compute.manager [None req-8ff32a28-977e-4c31-a0d5-5e0682d1957c tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.251514] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be256e3f-10ac-4bac-ad73-439cf2496486 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.329255] env[62521]: DEBUG oslo_vmware.api [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318652, 'name': PowerOffVM_Task, 'duration_secs': 0.22215} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.329549] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 876.329724] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 876.329987] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-baa0fa7e-6f02-4fce-9dfa-10a2e1068859 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.400204] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 876.400398] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 876.400584] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleting the datastore file [datastore2] acfd78aa-a51e-4a1c-b01a-15854fe0635c {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.400850] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92cbe7be-f819-439e-8059-56e9b076237a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.407692] env[62521]: DEBUG oslo_vmware.api [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 876.407692] env[62521]: value = "task-1318654" [ 876.407692] env[62521]: _type = "Task" [ 876.407692] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.415273] env[62521]: DEBUG oslo_vmware.api [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318654, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.454746] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "refresh_cache-d574f077-b221-4f91-8b54-0915421cb36f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.455178] env[62521]: DEBUG nova.compute.manager [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Instance network_info: |[{"id": "862343de-1cfa-4880-832a-bc8881cd0c30", "address": "fa:16:3e:d5:4b:bc", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap862343de-1c", "ovs_interfaceid": "862343de-1cfa-4880-832a-bc8881cd0c30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 876.455734] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:4b:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '84aee122-f630-43c5-9cc1-3a38d3819c82', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '862343de-1cfa-4880-832a-bc8881cd0c30', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.467961] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Creating folder: Project (c533bd4917b4466e9c5c76a9caf8e807). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 876.468537] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3845c62a-8d1d-4244-a1e8-4f7d7a2dce30 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.480841] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Created folder: Project (c533bd4917b4466e9c5c76a9caf8e807) in parent group-v282025. [ 876.480841] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Creating folder: Instances. Parent ref: group-v282110. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 876.481214] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ed1b5c8f-5023-4f9c-bc09-49eb94317f8c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.492542] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Created folder: Instances in parent group-v282110. [ 876.492931] env[62521]: DEBUG oslo.service.loopingcall [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.493200] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.494180] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dcd55cfd-44df-406b-8d09-1bd0f9b11a16 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.518055] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.518055] env[62521]: value = "task-1318657" [ 876.518055] env[62521]: _type = "Task" [ 876.518055] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.527646] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318657, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.532686] env[62521]: DEBUG nova.compute.manager [req-41ac82d9-e58d-46d6-8dfc-09efca1c60e8 req-c8584f4f-4851-4bf0-9ad0-87d43576401a service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Received event network-changed-862343de-1cfa-4880-832a-bc8881cd0c30 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.532898] env[62521]: DEBUG nova.compute.manager [req-41ac82d9-e58d-46d6-8dfc-09efca1c60e8 req-c8584f4f-4851-4bf0-9ad0-87d43576401a service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Refreshing instance network info cache due to event network-changed-862343de-1cfa-4880-832a-bc8881cd0c30. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 876.533129] env[62521]: DEBUG oslo_concurrency.lockutils [req-41ac82d9-e58d-46d6-8dfc-09efca1c60e8 req-c8584f4f-4851-4bf0-9ad0-87d43576401a service nova] Acquiring lock "refresh_cache-d574f077-b221-4f91-8b54-0915421cb36f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.533279] env[62521]: DEBUG oslo_concurrency.lockutils [req-41ac82d9-e58d-46d6-8dfc-09efca1c60e8 req-c8584f4f-4851-4bf0-9ad0-87d43576401a service nova] Acquired lock "refresh_cache-d574f077-b221-4f91-8b54-0915421cb36f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.533444] env[62521]: DEBUG nova.network.neutron [req-41ac82d9-e58d-46d6-8dfc-09efca1c60e8 req-c8584f4f-4851-4bf0-9ad0-87d43576401a service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Refreshing network info cache for port 862343de-1cfa-4880-832a-bc8881cd0c30 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.585184] env[62521]: DEBUG nova.scheduler.client.report [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.732027] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.919752] env[62521]: DEBUG oslo_vmware.api [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318654, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.425086} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.919938] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 876.920266] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 876.921360] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 876.921360] env[62521]: INFO nova.compute.manager [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 876.921598] env[62521]: DEBUG oslo.service.loopingcall [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.921918] env[62521]: DEBUG nova.compute.manager [-] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 876.922123] env[62521]: DEBUG nova.network.neutron [-] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 877.028813] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318657, 'name': CreateVM_Task, 'duration_secs': 0.398531} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.030277] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.030277] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.030277] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.030749] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.031124] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9958c479-3bf9-44c6-9fca-046f61f13e5c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.039831] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 877.039831] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52170444-01a9-00ca-a56b-e6afab15e70c" [ 877.039831] env[62521]: _type = "Task" [ 877.039831] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.056909] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52170444-01a9-00ca-a56b-e6afab15e70c, 'name': SearchDatastore_Task, 'duration_secs': 0.009974} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.057362] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.057705] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.058068] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.058347] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.058523] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.058795] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b2c516f-05c4-4f24-b5c9-933c83ff9bc5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.067503] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.067503] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.068237] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1cecdfb-d7d4-429a-af66-5c324076a435 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.073324] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 877.073324] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529ef884-fbfa-ee9b-2f68-d5186b029824" [ 877.073324] env[62521]: _type = "Task" [ 877.073324] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.081121] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529ef884-fbfa-ee9b-2f68-d5186b029824, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.092096] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.092250] env[62521]: DEBUG nova.compute.manager [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.094801] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.800s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.096258] env[62521]: INFO nova.compute.claims [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.396788] env[62521]: DEBUG nova.network.neutron [req-41ac82d9-e58d-46d6-8dfc-09efca1c60e8 req-c8584f4f-4851-4bf0-9ad0-87d43576401a service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Updated VIF entry in instance network info cache for port 862343de-1cfa-4880-832a-bc8881cd0c30. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 877.397251] env[62521]: DEBUG nova.network.neutron [req-41ac82d9-e58d-46d6-8dfc-09efca1c60e8 req-c8584f4f-4851-4bf0-9ad0-87d43576401a service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Updating instance_info_cache with network_info: [{"id": "862343de-1cfa-4880-832a-bc8881cd0c30", "address": "fa:16:3e:d5:4b:bc", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap862343de-1c", "ovs_interfaceid": "862343de-1cfa-4880-832a-bc8881cd0c30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.583782] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529ef884-fbfa-ee9b-2f68-d5186b029824, 'name': SearchDatastore_Task, 'duration_secs': 0.008391} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.584944] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-460bb73f-b48f-431f-89fe-86045d377973 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.590507] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 877.590507] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52044b7e-5982-2fc0-9e6f-ea1aece1849d" [ 877.590507] env[62521]: _type = "Task" [ 877.590507] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.598621] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52044b7e-5982-2fc0-9e6f-ea1aece1849d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.600971] env[62521]: DEBUG nova.compute.utils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.604235] env[62521]: DEBUG nova.compute.manager [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 877.604825] env[62521]: DEBUG nova.network.neutron [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 877.699414] env[62521]: DEBUG nova.policy [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0e99fc3ead94078b185bc599ff247ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25420ed4f82e478989a011d40f037a24', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 877.730779] env[62521]: DEBUG nova.network.neutron [-] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.766875] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 877.767177] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282109', 'volume_id': '5e701085-e961-46c7-8528-8e06b1899590', 'name': 'volume-5e701085-e961-46c7-8528-8e06b1899590', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af91de53-2ef8-48d1-b8c1-7bd0809f4ae4', 'attached_at': '', 'detached_at': '', 'volume_id': '5e701085-e961-46c7-8528-8e06b1899590', 'serial': '5e701085-e961-46c7-8528-8e06b1899590'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 877.768011] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250a0761-d0df-4607-93a5-0f11c3905427 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.784789] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5247d6df-ceba-4643-81e6-52a175a09a5f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.809318] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] volume-5e701085-e961-46c7-8528-8e06b1899590/volume-5e701085-e961-46c7-8528-8e06b1899590.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.810023] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-746e1023-cc65-4f92-9df0-a18b5085f261 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.827501] env[62521]: DEBUG oslo_vmware.api [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 877.827501] env[62521]: value = "task-1318658" [ 877.827501] env[62521]: _type = "Task" [ 877.827501] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.835527] env[62521]: DEBUG oslo_vmware.api [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318658, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.900034] env[62521]: DEBUG oslo_concurrency.lockutils [req-41ac82d9-e58d-46d6-8dfc-09efca1c60e8 req-c8584f4f-4851-4bf0-9ad0-87d43576401a service nova] Releasing lock "refresh_cache-d574f077-b221-4f91-8b54-0915421cb36f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.999139] env[62521]: DEBUG nova.network.neutron [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Successfully created port: 80906e15-3b2b-48f3-8c94-b2061848d775 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.102912] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52044b7e-5982-2fc0-9e6f-ea1aece1849d, 'name': SearchDatastore_Task, 'duration_secs': 0.028417} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.103192] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.103465] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] d574f077-b221-4f91-8b54-0915421cb36f/d574f077-b221-4f91-8b54-0915421cb36f.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.103733] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a871bd0-1cf3-4e47-8ed8-a4aa5058e8a8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.105870] env[62521]: DEBUG nova.compute.manager [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.116505] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 878.116505] env[62521]: value = "task-1318659" [ 878.116505] env[62521]: _type = "Task" [ 878.116505] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.124717] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318659, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.233046] env[62521]: INFO nova.compute.manager [-] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Took 1.31 seconds to deallocate network for instance. [ 878.342672] env[62521]: DEBUG oslo_vmware.api [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318658, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.388053] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f6ce05-9dc1-46ff-bc8d-7942030f7ded {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.396397] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa808c7-be87-4547-9cb6-043b53bec792 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.427418] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb183ea-4f79-4b80-bbe7-c2e60b8fd8b4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.436066] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cb5d49-5916-48a6-90f9-a8c91938c5c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.452090] env[62521]: DEBUG nova.compute.provider_tree [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.565360] env[62521]: DEBUG nova.compute.manager [req-ff89f369-6ff3-49fe-aa16-b5f10c49036a req-f42601d5-fd0c-4d02-adaa-3512697fca72 service nova] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Received event network-vif-deleted-9a829e1f-fed8-4e5c-a3c1-481e993b2551 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.632345] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318659, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.743119] env[62521]: DEBUG oslo_concurrency.lockutils [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.838852] env[62521]: DEBUG oslo_vmware.api [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318658, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.954699] env[62521]: DEBUG nova.scheduler.client.report [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.119041] env[62521]: DEBUG nova.compute.manager [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.130345] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318659, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554587} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.130615] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] d574f077-b221-4f91-8b54-0915421cb36f/d574f077-b221-4f91-8b54-0915421cb36f.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.131411] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.131411] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5fa5b58-a81d-459e-b159-68e2503e0882 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.138332] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 879.138332] env[62521]: value = "task-1318660" [ 879.138332] env[62521]: _type = "Task" [ 879.138332] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.146898] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318660, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.150299] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='6d00f28b952f638a279293ff94ba9d1e',container_format='bare',created_at=2024-10-20T00:35:25Z,direct_url=,disk_format='vmdk',id=9f64b520-b041-469a-8080-6b597c951678,min_disk=1,min_ram=0,name='tempest-test-snap-375265669',owner='25420ed4f82e478989a011d40f037a24',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-20T00:35:42Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.150558] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.150716] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.150904] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.151087] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.151257] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.151497] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.151665] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.151838] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.152013] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.152264] env[62521]: DEBUG nova.virt.hardware [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.153071] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d13e9d-5ebe-452e-8ee5-8f51222b4049 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.160445] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa9feab-11f3-4a4f-be75-49d6ee335ed2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.341747] env[62521]: DEBUG oslo_vmware.api [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318658, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.462266] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.462266] env[62521]: DEBUG nova.compute.manager [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.463685] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.564s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.613696] env[62521]: DEBUG nova.network.neutron [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Successfully updated port: 80906e15-3b2b-48f3-8c94-b2061848d775 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.648423] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318660, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.270263} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.648706] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.649535] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df50450c-1aff-4e76-9549-69e8292f0126 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.674978] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] d574f077-b221-4f91-8b54-0915421cb36f/d574f077-b221-4f91-8b54-0915421cb36f.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.677837] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-539cfe0f-ddac-4711-888f-2f6fd0782844 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.698946] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 879.698946] env[62521]: value = "task-1318661" [ 879.698946] env[62521]: _type = "Task" [ 879.698946] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.712485] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318661, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.734780] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d509ea81-f9f8-4715-83a7-0c7ed735a0d2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.742031] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86797e91-e0c3-4c58-b9e4-745b8df14117 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.784526] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbee5b2-67c8-4d47-a56d-18d5478f4025 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.793384] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660b672a-e2dd-4274-8d95-e100199e9c61 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.809076] env[62521]: DEBUG nova.compute.provider_tree [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.840821] env[62521]: DEBUG oslo_vmware.api [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318658, 'name': ReconfigVM_Task, 'duration_secs': 1.87413} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.841261] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Reconfigured VM instance instance-0000003a to attach disk [datastore2] volume-5e701085-e961-46c7-8528-8e06b1899590/volume-5e701085-e961-46c7-8528-8e06b1899590.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.846903] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14a41c4f-a1a5-4c01-b995-84f7f80a1de7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.864018] env[62521]: DEBUG oslo_vmware.api [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 879.864018] env[62521]: value = "task-1318662" [ 879.864018] env[62521]: _type = "Task" [ 879.864018] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.871409] env[62521]: DEBUG oslo_vmware.api [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318662, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.969107] env[62521]: DEBUG nova.compute.utils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.970590] env[62521]: DEBUG nova.compute.manager [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.971216] env[62521]: DEBUG nova.network.neutron [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.062742] env[62521]: DEBUG nova.policy [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c63ec03cec5437285d6a99fc2a1d791', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '840dfca13260497b93990e4875cfd532', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.120037] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "refresh_cache-1d1d2235-748a-4f27-a915-b69f9110257b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.120037] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "refresh_cache-1d1d2235-748a-4f27-a915-b69f9110257b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.120037] env[62521]: DEBUG nova.network.neutron [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.213409] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318661, 'name': ReconfigVM_Task, 'duration_secs': 0.311268} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.213532] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Reconfigured VM instance instance-00000047 to attach disk [datastore2] d574f077-b221-4f91-8b54-0915421cb36f/d574f077-b221-4f91-8b54-0915421cb36f.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.214192] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-670cbfeb-1972-41c9-8728-2dd0ea3c85cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.220955] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 880.220955] env[62521]: value = "task-1318663" [ 880.220955] env[62521]: _type = "Task" [ 880.220955] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.231433] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318663, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.317247] env[62521]: DEBUG nova.scheduler.client.report [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.372150] env[62521]: DEBUG oslo_vmware.api [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318662, 'name': ReconfigVM_Task, 'duration_secs': 0.142026} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.372475] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282109', 'volume_id': '5e701085-e961-46c7-8528-8e06b1899590', 'name': 'volume-5e701085-e961-46c7-8528-8e06b1899590', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af91de53-2ef8-48d1-b8c1-7bd0809f4ae4', 'attached_at': '', 'detached_at': '', 'volume_id': '5e701085-e961-46c7-8528-8e06b1899590', 'serial': '5e701085-e961-46c7-8528-8e06b1899590'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 880.474322] env[62521]: DEBUG nova.compute.manager [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 880.502630] env[62521]: DEBUG nova.network.neutron [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Successfully created port: ecf8506b-ea0e-48e2-8868-af841b45fe97 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.597734] env[62521]: DEBUG nova.compute.manager [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Received event network-vif-plugged-80906e15-3b2b-48f3-8c94-b2061848d775 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.597986] env[62521]: DEBUG oslo_concurrency.lockutils [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] Acquiring lock "1d1d2235-748a-4f27-a915-b69f9110257b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.598215] env[62521]: DEBUG oslo_concurrency.lockutils [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] Lock "1d1d2235-748a-4f27-a915-b69f9110257b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.598404] env[62521]: DEBUG oslo_concurrency.lockutils [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] Lock "1d1d2235-748a-4f27-a915-b69f9110257b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.598831] env[62521]: DEBUG nova.compute.manager [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] No waiting events found dispatching network-vif-plugged-80906e15-3b2b-48f3-8c94-b2061848d775 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.599032] env[62521]: WARNING nova.compute.manager [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Received unexpected event network-vif-plugged-80906e15-3b2b-48f3-8c94-b2061848d775 for instance with vm_state building and task_state spawning. [ 880.599206] env[62521]: DEBUG nova.compute.manager [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Received event network-changed-80906e15-3b2b-48f3-8c94-b2061848d775 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.599365] env[62521]: DEBUG nova.compute.manager [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Refreshing instance network info cache due to event network-changed-80906e15-3b2b-48f3-8c94-b2061848d775. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.600187] env[62521]: DEBUG oslo_concurrency.lockutils [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] Acquiring lock "refresh_cache-1d1d2235-748a-4f27-a915-b69f9110257b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.677075] env[62521]: DEBUG nova.network.neutron [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 880.732314] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318663, 'name': Rename_Task, 'duration_secs': 0.164582} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.732609] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.733703] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5090d16f-641f-41bc-b9a8-2db368582a2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.739877] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 880.739877] env[62521]: value = "task-1318664" [ 880.739877] env[62521]: _type = "Task" [ 880.739877] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.748855] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.822962] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.359s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.823229] env[62521]: INFO nova.compute.manager [None req-fcae5a1a-74e6-4ac6-904e-b8062b3c62b5 tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Successfully reverted task state from rebuilding on failure for instance. [ 880.829644] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.822s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.829898] env[62521]: DEBUG nova.objects.instance [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lazy-loading 'resources' on Instance uuid 2a0ccda4-87a6-4bb3-bf34-4434faf574e6 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.072021] env[62521]: DEBUG nova.network.neutron [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Updating instance_info_cache with network_info: [{"id": "80906e15-3b2b-48f3-8c94-b2061848d775", "address": "fa:16:3e:5d:52:21", "network": {"id": "ce7ccdc7-e99d-48bd-93a9-c8c14725d58b", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1698306226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25420ed4f82e478989a011d40f037a24", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b56036cd-97ac-47f5-9089-7b38bfe99228", "external-id": "nsx-vlan-transportzone-301", "segmentation_id": 301, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80906e15-3b", "ovs_interfaceid": "80906e15-3b2b-48f3-8c94-b2061848d775", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.250315] env[62521]: DEBUG oslo_vmware.api [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318664, 'name': PowerOnVM_Task, 'duration_secs': 0.471056} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.250563] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 881.250770] env[62521]: INFO nova.compute.manager [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Took 7.39 seconds to spawn the instance on the hypervisor. [ 881.250948] env[62521]: DEBUG nova.compute.manager [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.251781] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393c461d-8a48-4936-a05c-3c8db1af61ea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.421726] env[62521]: DEBUG nova.objects.instance [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lazy-loading 'flavor' on Instance uuid af91de53-2ef8-48d1-b8c1-7bd0809f4ae4 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.483779] env[62521]: DEBUG nova.compute.manager [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.527272] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.527621] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.527792] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.527981] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.528147] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.528304] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.528551] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.528723] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.528895] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.529074] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.529260] env[62521]: DEBUG nova.virt.hardware [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.530157] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560a6200-5850-4360-a7ea-acba643c2705 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.541123] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6711a198-4941-45df-85b1-1d95d7aba31b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.575197] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "refresh_cache-1d1d2235-748a-4f27-a915-b69f9110257b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.575568] env[62521]: DEBUG nova.compute.manager [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Instance network_info: |[{"id": "80906e15-3b2b-48f3-8c94-b2061848d775", "address": "fa:16:3e:5d:52:21", "network": {"id": "ce7ccdc7-e99d-48bd-93a9-c8c14725d58b", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1698306226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25420ed4f82e478989a011d40f037a24", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b56036cd-97ac-47f5-9089-7b38bfe99228", "external-id": "nsx-vlan-transportzone-301", "segmentation_id": 301, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80906e15-3b", "ovs_interfaceid": "80906e15-3b2b-48f3-8c94-b2061848d775", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 881.575888] env[62521]: DEBUG oslo_concurrency.lockutils [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] Acquired lock "refresh_cache-1d1d2235-748a-4f27-a915-b69f9110257b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.576170] env[62521]: DEBUG nova.network.neutron [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Refreshing network info cache for port 80906e15-3b2b-48f3-8c94-b2061848d775 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.577666] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:52:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b56036cd-97ac-47f5-9089-7b38bfe99228', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '80906e15-3b2b-48f3-8c94-b2061848d775', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.589493] env[62521]: DEBUG oslo.service.loopingcall [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.593307] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.594043] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b00603ee-f026-4372-b2d4-f37eeb5e6aa3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.622847] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.622847] env[62521]: value = "task-1318665" [ 881.622847] env[62521]: _type = "Task" [ 881.622847] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.635608] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318665, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.668611] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af0fbc8-820a-4540-958a-652363f202ea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.674946] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d81dd96-64e0-480b-b9f9-30fbd7ca0851 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.717179] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392dcae2-55bd-4032-940f-565e794eb250 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.725691] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f367717-563f-4b00-920b-7573efbbf821 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.739771] env[62521]: DEBUG nova.compute.provider_tree [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.774466] env[62521]: INFO nova.compute.manager [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Took 23.70 seconds to build instance. [ 881.931512] env[62521]: DEBUG oslo_concurrency.lockutils [None req-34d80364-9ab9-4569-9cc3-f449d1b55b7b tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.769s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.094887] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.095294] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.140288] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318665, 'name': CreateVM_Task, 'duration_secs': 0.501974} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.142528] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.143241] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.143418] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.143805] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.144370] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-226bffcc-9066-4451-9dbf-34f22f744974 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.149166] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 882.149166] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a39881-d56b-aceb-6d4f-8752410d89e3" [ 882.149166] env[62521]: _type = "Task" [ 882.149166] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.156966] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a39881-d56b-aceb-6d4f-8752410d89e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.243344] env[62521]: DEBUG nova.scheduler.client.report [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.276468] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d8aa848-7310-48f7-bad3-511343c6c2d0 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "d574f077-b221-4f91-8b54-0915421cb36f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.213s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.598683] env[62521]: DEBUG nova.compute.manager [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 882.665969] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.666323] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Processing image 9f64b520-b041-469a-8080-6b597c951678 {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.666637] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678/9f64b520-b041-469a-8080-6b597c951678.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.666794] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678/9f64b520-b041-469a-8080-6b597c951678.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.667026] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.667509] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e851ebe2-9d85-426c-84d6-e3f4de7c241c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.686805] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.687043] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.687779] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8d051f4-7f13-4062-8f7d-1d6a779357e8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.696033] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 882.696033] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d9c689-1f80-6df2-3966-685cf8f8f2e3" [ 882.696033] env[62521]: _type = "Task" [ 882.696033] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.703861] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d9c689-1f80-6df2-3966-685cf8f8f2e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.748482] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.919s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.750883] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.560s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.752890] env[62521]: INFO nova.compute.claims [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.792781] env[62521]: INFO nova.scheduler.client.report [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Deleted allocations for instance 2a0ccda4-87a6-4bb3-bf34-4434faf574e6 [ 882.966858] env[62521]: DEBUG nova.network.neutron [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Updated VIF entry in instance network info cache for port 80906e15-3b2b-48f3-8c94-b2061848d775. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.967272] env[62521]: DEBUG nova.network.neutron [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Updating instance_info_cache with network_info: [{"id": "80906e15-3b2b-48f3-8c94-b2061848d775", "address": "fa:16:3e:5d:52:21", "network": {"id": "ce7ccdc7-e99d-48bd-93a9-c8c14725d58b", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1698306226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25420ed4f82e478989a011d40f037a24", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b56036cd-97ac-47f5-9089-7b38bfe99228", "external-id": "nsx-vlan-transportzone-301", "segmentation_id": 301, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80906e15-3b", "ovs_interfaceid": "80906e15-3b2b-48f3-8c94-b2061848d775", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.127516] env[62521]: DEBUG nova.compute.manager [req-8b37b0f2-c132-4a3c-8eef-d92136fd2922 req-19b68dc9-b593-4d9b-a207-8850cababfd2 service nova] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Received event network-vif-plugged-ecf8506b-ea0e-48e2-8868-af841b45fe97 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.127886] env[62521]: DEBUG oslo_concurrency.lockutils [req-8b37b0f2-c132-4a3c-8eef-d92136fd2922 req-19b68dc9-b593-4d9b-a207-8850cababfd2 service nova] Acquiring lock "3447712d-daa6-4329-8015-6474cca1a5e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.127947] env[62521]: DEBUG oslo_concurrency.lockutils [req-8b37b0f2-c132-4a3c-8eef-d92136fd2922 req-19b68dc9-b593-4d9b-a207-8850cababfd2 service nova] Lock "3447712d-daa6-4329-8015-6474cca1a5e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.128146] env[62521]: DEBUG oslo_concurrency.lockutils [req-8b37b0f2-c132-4a3c-8eef-d92136fd2922 req-19b68dc9-b593-4d9b-a207-8850cababfd2 service nova] Lock "3447712d-daa6-4329-8015-6474cca1a5e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.128291] env[62521]: DEBUG nova.compute.manager [req-8b37b0f2-c132-4a3c-8eef-d92136fd2922 req-19b68dc9-b593-4d9b-a207-8850cababfd2 service nova] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] No waiting events found dispatching network-vif-plugged-ecf8506b-ea0e-48e2-8868-af841b45fe97 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 883.128452] env[62521]: WARNING nova.compute.manager [req-8b37b0f2-c132-4a3c-8eef-d92136fd2922 req-19b68dc9-b593-4d9b-a207-8850cababfd2 service nova] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Received unexpected event network-vif-plugged-ecf8506b-ea0e-48e2-8868-af841b45fe97 for instance with vm_state building and task_state spawning. [ 883.136050] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.208031] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Preparing fetch location {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 883.208285] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Fetch image to [datastore2] OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7/OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7.vmdk {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 883.208471] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Downloading stream optimized image 9f64b520-b041-469a-8080-6b597c951678 to [datastore2] OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7/OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7.vmdk on the data store datastore2 as vApp {{(pid=62521) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 883.208651] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Downloading image file data 9f64b520-b041-469a-8080-6b597c951678 to the ESX as VM named 'OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7' {{(pid=62521) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 883.211029] env[62521]: DEBUG nova.network.neutron [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Successfully updated port: ecf8506b-ea0e-48e2-8868-af841b45fe97 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.289968] env[62521]: DEBUG oslo_vmware.rw_handles [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 883.289968] env[62521]: value = "resgroup-9" [ 883.289968] env[62521]: _type = "ResourcePool" [ 883.289968] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 883.290397] env[62521]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-92d8b6aa-44c8-40c7-a080-8e59ec6b0bd4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.311091] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0c01276d-4967-4aab-9e21-eae24da217f9 tempest-ServerMetadataTestJSON-751018245 tempest-ServerMetadataTestJSON-751018245-project-member] Lock "2a0ccda4-87a6-4bb3-bf34-4434faf574e6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.298s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.318358] env[62521]: DEBUG oslo_vmware.rw_handles [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lease: (returnval){ [ 883.318358] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529e508e-8fc4-0dab-49e1-5e2f96f440a5" [ 883.318358] env[62521]: _type = "HttpNfcLease" [ 883.318358] env[62521]: } obtained for vApp import into resource pool (val){ [ 883.318358] env[62521]: value = "resgroup-9" [ 883.318358] env[62521]: _type = "ResourcePool" [ 883.318358] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 883.318693] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the lease: (returnval){ [ 883.318693] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529e508e-8fc4-0dab-49e1-5e2f96f440a5" [ 883.318693] env[62521]: _type = "HttpNfcLease" [ 883.318693] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 883.327272] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 883.327272] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529e508e-8fc4-0dab-49e1-5e2f96f440a5" [ 883.327272] env[62521]: _type = "HttpNfcLease" [ 883.327272] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 883.470360] env[62521]: DEBUG oslo_concurrency.lockutils [req-416d10e2-c55c-419d-868e-f5f7f221331a req-f67e1c63-7b41-4df6-88d1-9d5db13c8b93 service nova] Releasing lock "refresh_cache-1d1d2235-748a-4f27-a915-b69f9110257b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.719028] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "refresh_cache-3447712d-daa6-4329-8015-6474cca1a5e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.719028] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired lock "refresh_cache-3447712d-daa6-4329-8015-6474cca1a5e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.719028] env[62521]: DEBUG nova.network.neutron [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.833126] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 883.833126] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529e508e-8fc4-0dab-49e1-5e2f96f440a5" [ 883.833126] env[62521]: _type = "HttpNfcLease" [ 883.833126] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 884.169455] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c795f0-aaa3-46d6-9d57-cf1a22cdd01d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.179357] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c9ff58-ae15-45d4-80e3-816cc99aaf9b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.214296] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e179877a-d628-43ed-8b3a-75e602f27b7e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.221430] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b29819-c3dc-4af5-93ce-2b9db1ba7318 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.234019] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "563ad3b2-e53e-4682-9cde-bd6f709718de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.234308] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "563ad3b2-e53e-4682-9cde-bd6f709718de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.246951] env[62521]: DEBUG nova.compute.provider_tree [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.308442] env[62521]: DEBUG nova.network.neutron [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.327913] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 884.327913] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529e508e-8fc4-0dab-49e1-5e2f96f440a5" [ 884.327913] env[62521]: _type = "HttpNfcLease" [ 884.327913] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 884.480922] env[62521]: DEBUG nova.network.neutron [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Updating instance_info_cache with network_info: [{"id": "ecf8506b-ea0e-48e2-8868-af841b45fe97", "address": "fa:16:3e:3f:de:90", "network": {"id": "5b74eda2-668c-4d8f-b4d8-a5e748613ee1", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-574755447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "840dfca13260497b93990e4875cfd532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapecf8506b-ea", "ovs_interfaceid": "ecf8506b-ea0e-48e2-8868-af841b45fe97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.737497] env[62521]: DEBUG nova.compute.manager [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.741182] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "611462d2-7f57-44b0-ac36-db32af3d0dd0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.741514] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "611462d2-7f57-44b0-ac36-db32af3d0dd0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.752204] env[62521]: DEBUG nova.scheduler.client.report [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.828467] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 884.828467] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529e508e-8fc4-0dab-49e1-5e2f96f440a5" [ 884.828467] env[62521]: _type = "HttpNfcLease" [ 884.828467] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 884.828688] env[62521]: DEBUG oslo_vmware.rw_handles [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 884.828688] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529e508e-8fc4-0dab-49e1-5e2f96f440a5" [ 884.828688] env[62521]: _type = "HttpNfcLease" [ 884.828688] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 884.829428] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5e7fb3-667f-4f47-9a70-9f2ec8efbbb4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.837280] env[62521]: DEBUG oslo_vmware.rw_handles [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f34c2b-7323-f782-cf2d-801d4ef4c1c6/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 884.837280] env[62521]: DEBUG oslo_vmware.rw_handles [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f34c2b-7323-f782-cf2d-801d4ef4c1c6/disk-0.vmdk. {{(pid=62521) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 884.904073] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7b2029c7-7c60-4bae-8b4a-4e185311f53a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.984577] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Releasing lock "refresh_cache-3447712d-daa6-4329-8015-6474cca1a5e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.984916] env[62521]: DEBUG nova.compute.manager [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Instance network_info: |[{"id": "ecf8506b-ea0e-48e2-8868-af841b45fe97", "address": "fa:16:3e:3f:de:90", "network": {"id": "5b74eda2-668c-4d8f-b4d8-a5e748613ee1", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-574755447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "840dfca13260497b93990e4875cfd532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapecf8506b-ea", "ovs_interfaceid": "ecf8506b-ea0e-48e2-8868-af841b45fe97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 884.985382] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:de:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '415e68b4-3766-4359-afe2-f8563910d98c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ecf8506b-ea0e-48e2-8868-af841b45fe97', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.993280] env[62521]: DEBUG oslo.service.loopingcall [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.993534] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.993764] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35cf70d6-f6dd-49cc-9070-b89fe219a353 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.018578] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.018578] env[62521]: value = "task-1318667" [ 885.018578] env[62521]: _type = "Task" [ 885.018578] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.028949] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318667, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.158591] env[62521]: DEBUG nova.compute.manager [req-6644d6d0-65c2-4690-a07e-3128d3a67316 req-0a68633f-148b-4761-9b26-b0e6ad49e23c service nova] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Received event network-changed-ecf8506b-ea0e-48e2-8868-af841b45fe97 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.158894] env[62521]: DEBUG nova.compute.manager [req-6644d6d0-65c2-4690-a07e-3128d3a67316 req-0a68633f-148b-4761-9b26-b0e6ad49e23c service nova] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Refreshing instance network info cache due to event network-changed-ecf8506b-ea0e-48e2-8868-af841b45fe97. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.159328] env[62521]: DEBUG oslo_concurrency.lockutils [req-6644d6d0-65c2-4690-a07e-3128d3a67316 req-0a68633f-148b-4761-9b26-b0e6ad49e23c service nova] Acquiring lock "refresh_cache-3447712d-daa6-4329-8015-6474cca1a5e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.159536] env[62521]: DEBUG oslo_concurrency.lockutils [req-6644d6d0-65c2-4690-a07e-3128d3a67316 req-0a68633f-148b-4761-9b26-b0e6ad49e23c service nova] Acquired lock "refresh_cache-3447712d-daa6-4329-8015-6474cca1a5e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.159733] env[62521]: DEBUG nova.network.neutron [req-6644d6d0-65c2-4690-a07e-3128d3a67316 req-0a68633f-148b-4761-9b26-b0e6ad49e23c service nova] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Refreshing network info cache for port ecf8506b-ea0e-48e2-8868-af841b45fe97 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.248123] env[62521]: DEBUG nova.compute.manager [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 885.260872] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.261506] env[62521]: DEBUG nova.compute.manager [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.264576] env[62521]: DEBUG oslo_concurrency.lockutils [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.839s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.265054] env[62521]: DEBUG nova.objects.instance [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lazy-loading 'resources' on Instance uuid e8cabcdc-d728-434d-a40d-fe751db086c3 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.275839] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.529616] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318667, 'name': CreateVM_Task, 'duration_secs': 0.358977} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.529781] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.530495] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.530672] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.531053] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.532784] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db8a96bf-2460-45ec-8c1c-1edb62d19125 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.537800] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 885.537800] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5201f8d2-09d1-03c3-43d6-ae6333919e00" [ 885.537800] env[62521]: _type = "Task" [ 885.537800] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.547752] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5201f8d2-09d1-03c3-43d6-ae6333919e00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.772886] env[62521]: DEBUG nova.compute.utils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.781128] env[62521]: DEBUG oslo_vmware.rw_handles [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Completed reading data from the image iterator. {{(pid=62521) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 885.781360] env[62521]: DEBUG oslo_vmware.rw_handles [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f34c2b-7323-f782-cf2d-801d4ef4c1c6/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 885.781698] env[62521]: DEBUG nova.compute.manager [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Not allocating networking since 'none' was specified. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 885.782602] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25372fa8-a68d-4a10-bdde-550f53272fd2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.790092] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.795553] env[62521]: DEBUG oslo_vmware.rw_handles [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f34c2b-7323-f782-cf2d-801d4ef4c1c6/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 885.795931] env[62521]: DEBUG oslo_vmware.rw_handles [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f34c2b-7323-f782-cf2d-801d4ef4c1c6/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 885.797216] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-46017f9d-8d72-4b80-961f-d154f9a3c9b6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.989940] env[62521]: DEBUG nova.network.neutron [req-6644d6d0-65c2-4690-a07e-3128d3a67316 req-0a68633f-148b-4761-9b26-b0e6ad49e23c service nova] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Updated VIF entry in instance network info cache for port ecf8506b-ea0e-48e2-8868-af841b45fe97. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 885.990262] env[62521]: DEBUG nova.network.neutron [req-6644d6d0-65c2-4690-a07e-3128d3a67316 req-0a68633f-148b-4761-9b26-b0e6ad49e23c service nova] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Updating instance_info_cache with network_info: [{"id": "ecf8506b-ea0e-48e2-8868-af841b45fe97", "address": "fa:16:3e:3f:de:90", "network": {"id": "5b74eda2-668c-4d8f-b4d8-a5e748613ee1", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-574755447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "840dfca13260497b93990e4875cfd532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapecf8506b-ea", "ovs_interfaceid": "ecf8506b-ea0e-48e2-8868-af841b45fe97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.992744] env[62521]: DEBUG oslo_vmware.rw_handles [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f34c2b-7323-f782-cf2d-801d4ef4c1c6/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 885.994032] env[62521]: INFO nova.virt.vmwareapi.images [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Downloaded image file data 9f64b520-b041-469a-8080-6b597c951678 [ 885.994032] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4036b7-1c78-47ad-8fc4-c802294f6c0a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.010995] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e01aa457-ff93-42e6-830d-884ddd7628ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.049860] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5201f8d2-09d1-03c3-43d6-ae6333919e00, 'name': SearchDatastore_Task, 'duration_secs': 0.011083} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.051049] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.051307] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.051559] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.051716] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.051902] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.052700] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06053864-f3de-4bec-bead-1b1522eb4b93 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.055226] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3b490a2-7cbf-4956-a216-fd6b2d652d47 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.061732] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbf7975-2b3a-486d-8d06-4639f3a5fb03 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.065530] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.065735] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 886.066814] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef4bb393-d7ae-40ac-8bb3-3b94681ef0e1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.096316] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3caf076f-ee69-46a6-8379-fd13b1596d87 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.099352] env[62521]: INFO nova.virt.vmwareapi.images [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] The imported VM was unregistered [ 886.102134] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Caching image {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 886.102134] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Creating directory with path [datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678 {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.102134] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 886.102134] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a20cc0-2e77-2050-808f-4e8b6d75984b" [ 886.102134] env[62521]: _type = "Task" [ 886.102134] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.102134] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b63e0b13-2954-4496-ac79-c8ce690ce422 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.109724] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c12bdcb-da33-4eb9-b354-a1889588cf5c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.116449] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a20cc0-2e77-2050-808f-4e8b6d75984b, 'name': SearchDatastore_Task, 'duration_secs': 0.022551} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.117419] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beb7870e-d3a0-4bd0-8712-76de0f721401 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.126881] env[62521]: DEBUG nova.compute.provider_tree [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.129375] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Created directory with path [datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678 {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.129588] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7/OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7.vmdk to [datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678/9f64b520-b041-469a-8080-6b597c951678.vmdk. {{(pid=62521) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 886.130015] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-44f1cfc6-a4d5-4479-95d2-b4f5924750b1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.132922] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 886.132922] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cf166e-23cc-91d0-b5a3-2114014e7f6e" [ 886.132922] env[62521]: _type = "Task" [ 886.132922] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.137296] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 886.137296] env[62521]: value = "task-1318669" [ 886.137296] env[62521]: _type = "Task" [ 886.137296] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.143493] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cf166e-23cc-91d0-b5a3-2114014e7f6e, 'name': SearchDatastore_Task, 'duration_secs': 0.008628} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.144609] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.144881] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 3447712d-daa6-4329-8015-6474cca1a5e4/3447712d-daa6-4329-8015-6474cca1a5e4.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.145147] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-294483aa-2505-48a3-85fe-3671bf7677a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.150477] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318669, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.154807] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 886.154807] env[62521]: value = "task-1318670" [ 886.154807] env[62521]: _type = "Task" [ 886.154807] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.162418] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318670, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.288338] env[62521]: DEBUG nova.compute.manager [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.494460] env[62521]: DEBUG oslo_concurrency.lockutils [req-6644d6d0-65c2-4690-a07e-3128d3a67316 req-0a68633f-148b-4761-9b26-b0e6ad49e23c service nova] Releasing lock "refresh_cache-3447712d-daa6-4329-8015-6474cca1a5e4" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.632230] env[62521]: DEBUG nova.scheduler.client.report [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.648430] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318669, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.665175] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318670, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.138530] env[62521]: DEBUG oslo_concurrency.lockutils [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.874s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.141060] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.930s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.141218] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.141337] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 887.141708] env[62521]: DEBUG oslo_concurrency.lockutils [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 13.180s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.143691] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c800a6db-b969-437c-b2b3-57852c740900 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.166530] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318669, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.168366] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdc4463-c033-4c11-ae2e-c141f392f3c3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.177689] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318670, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.186630] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410808ef-6801-4afa-bf85-f8b4c5403b6f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.193869] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be16a260-20d6-4ea1-b75e-293ba54526bb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.224124] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180750MB free_disk=133GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 887.224316] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.298655] env[62521]: DEBUG nova.compute.manager [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.326010] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.326366] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.326537] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.326730] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.326881] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.327077] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.327342] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.327537] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.327645] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.327816] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.327990] env[62521]: DEBUG nova.virt.hardware [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.328929] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e7c31f-5461-4172-8287-936e7705d828 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.337363] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b3a7b8-3a0d-45a2-9e90-cbaeacf4f876 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.351865] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.358363] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Creating folder: Project (d3978c74f3744a6f9e28b677f6621cfa). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 887.359170] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-620433e9-420e-4644-8699-4c239980de49 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.369966] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Created folder: Project (d3978c74f3744a6f9e28b677f6621cfa) in parent group-v282025. [ 887.370176] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Creating folder: Instances. Parent ref: group-v282116. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 887.370441] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c41c28a1-6b30-4909-9d22-9b4fa9ada57d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.381678] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Created folder: Instances in parent group-v282116. [ 887.381956] env[62521]: DEBUG oslo.service.loopingcall [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.382183] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 887.382400] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ab204c7-723b-4916-9504-ce5f1af0b8b8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.401788] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.401788] env[62521]: value = "task-1318673" [ 887.401788] env[62521]: _type = "Task" [ 887.401788] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.409686] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318673, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.656009] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318669, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.671748] env[62521]: DEBUG oslo_concurrency.lockutils [None req-156c8ab2-a3b1-4b2c-a76a-2dd065ad524d tempest-ServerActionsV293TestJSON-2139287260 tempest-ServerActionsV293TestJSON-2139287260-project-member] Lock "e8cabcdc-d728-434d-a40d-fe751db086c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.314s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.673661] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318670, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.08607} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.673990] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 3447712d-daa6-4329-8015-6474cca1a5e4/3447712d-daa6-4329-8015-6474cca1a5e4.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 887.674221] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.674551] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37d8ef8f-bae4-4ff8-b786-896b64013a18 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.684759] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 887.684759] env[62521]: value = "task-1318674" [ 887.684759] env[62521]: _type = "Task" [ 887.684759] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.700794] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318674, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.916136] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318673, 'name': CreateVM_Task, 'duration_secs': 0.377609} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.916303] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.916767] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.916942] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.917466] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.917843] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c61ada1-33d3-464b-99c8-575ca0d6623d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.921565] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98a69f6c-aaab-480f-87ee-4131d2c4bd13 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.925309] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 887.925309] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a59dfe-118a-ede6-c5ff-ece888c7ab2c" [ 887.925309] env[62521]: _type = "Task" [ 887.925309] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.932108] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c4cc7a-05a8-4727-a4a2-cea5f9704eab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.938925] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a59dfe-118a-ede6-c5ff-ece888c7ab2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.966936] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3118ec7f-be25-4ad4-8a84-fa0a9360abe5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.974907] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8501cd80-9320-4f40-b50e-fe81c6fd4aab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.990644] env[62521]: DEBUG nova.compute.provider_tree [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.153484] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318669, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.194339] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318674, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080974} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.194591] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.195439] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ae3180-bacd-4d45-9972-d0735068b0dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.218570] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 3447712d-daa6-4329-8015-6474cca1a5e4/3447712d-daa6-4329-8015-6474cca1a5e4.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.218921] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e11d7aa0-0f16-496e-ad62-c42cea3dbcd2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.239384] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 888.239384] env[62521]: value = "task-1318675" [ 888.239384] env[62521]: _type = "Task" [ 888.239384] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.248095] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318675, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.438598] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a59dfe-118a-ede6-c5ff-ece888c7ab2c, 'name': SearchDatastore_Task, 'duration_secs': 0.011884} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.439037] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.439485] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.439661] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.439866] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.440139] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.440592] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8bc1c61-f1bd-4863-b286-a8581ee8f9d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.452308] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.452618] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.453751] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73970351-2393-4e2e-b7e8-31e9f782f82f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.461230] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 888.461230] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f564c6-c368-4c8f-18a8-74fd227f2ca6" [ 888.461230] env[62521]: _type = "Task" [ 888.461230] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.473530] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f564c6-c368-4c8f-18a8-74fd227f2ca6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.494164] env[62521]: DEBUG nova.scheduler.client.report [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.655472] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318669, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.748715] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318675, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.977268] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f564c6-c368-4c8f-18a8-74fd227f2ca6, 'name': SearchDatastore_Task, 'duration_secs': 0.041621} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.978400] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf39c497-164f-4442-896c-faf7f0cf09af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.985898] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 888.985898] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5269f381-7c60-c4bc-17a3-0b70c4ff045d" [ 888.985898] env[62521]: _type = "Task" [ 888.985898] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.994359] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5269f381-7c60-c4bc-17a3-0b70c4ff045d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.158704] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318669, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.590058} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.158999] env[62521]: INFO nova.virt.vmwareapi.ds_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7/OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7.vmdk to [datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678/9f64b520-b041-469a-8080-6b597c951678.vmdk. [ 889.159228] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Cleaning up location [datastore2] OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7 {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 889.159394] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_9a2ec704-c803-4e3b-83e8-de93b75327d7 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.159652] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb42a36f-c2e0-4a01-bbe1-835ca8b56be6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.167213] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 889.167213] env[62521]: value = "task-1318676" [ 889.167213] env[62521]: _type = "Task" [ 889.167213] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.177140] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318676, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.249275] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318675, 'name': ReconfigVM_Task, 'duration_secs': 0.746908} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.249534] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 3447712d-daa6-4329-8015-6474cca1a5e4/3447712d-daa6-4329-8015-6474cca1a5e4.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.250223] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bdf31d74-9a9e-407e-9861-a2978eedb514 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.256293] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 889.256293] env[62521]: value = "task-1318677" [ 889.256293] env[62521]: _type = "Task" [ 889.256293] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.263773] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318677, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.496704] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5269f381-7c60-c4bc-17a3-0b70c4ff045d, 'name': SearchDatastore_Task, 'duration_secs': 0.009235} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.496989] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.497282] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] e2104c11-7713-4542-83da-bc4d2534113a/e2104c11-7713-4542-83da-bc4d2534113a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.497548] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c6f0040-deed-4c7b-853e-92ba96394c30 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.503670] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 889.503670] env[62521]: value = "task-1318678" [ 889.503670] env[62521]: _type = "Task" [ 889.503670] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.504771] env[62521]: DEBUG oslo_concurrency.lockutils [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.363s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.507709] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.404s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.509055] env[62521]: INFO nova.compute.claims [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.519827] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318678, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.677471] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318676, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.044998} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.677827] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.678093] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678/9f64b520-b041-469a-8080-6b597c951678.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.678437] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678/9f64b520-b041-469a-8080-6b597c951678.vmdk to [datastore2] 1d1d2235-748a-4f27-a915-b69f9110257b/1d1d2235-748a-4f27-a915-b69f9110257b.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.678713] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d93a5a53-4801-4e35-b512-12ad96af262c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.684932] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 889.684932] env[62521]: value = "task-1318679" [ 889.684932] env[62521]: _type = "Task" [ 889.684932] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.692503] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318679, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.766984] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318677, 'name': Rename_Task, 'duration_secs': 0.137786} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.767394] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 889.767632] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3db6943d-666a-4705-b526-987283b338bf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.774663] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 889.774663] env[62521]: value = "task-1318680" [ 889.774663] env[62521]: _type = "Task" [ 889.774663] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.783676] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318680, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.024046] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318678, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.091466] env[62521]: INFO nova.scheduler.client.report [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Deleted allocation for migration 29ec3101-66e2-4662-a2ab-f405aeb20552 [ 890.195442] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318679, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.285343] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318680, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.518050] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318678, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628172} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.518512] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] e2104c11-7713-4542-83da-bc4d2534113a/e2104c11-7713-4542-83da-bc4d2534113a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 890.518623] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.518906] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-32916b84-e1d9-4645-b134-a341016c6223 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.528330] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 890.528330] env[62521]: value = "task-1318681" [ 890.528330] env[62521]: _type = "Task" [ 890.528330] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.539270] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318681, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.599282] env[62521]: DEBUG oslo_concurrency.lockutils [None req-13f97356-62ce-46f7-aeed-33949b026b65 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 20.417s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.696014] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318679, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.785395] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318680, 'name': PowerOnVM_Task} progress is 76%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.787631] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae007ba-ff7a-4707-9c9b-d972497cec22 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.795577] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b26b93f-9030-4fbb-ad4c-f414a7a3f975 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.826220] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aef352a-6371-40f7-8580-7d657e4c2e49 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.834290] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3e8749-a167-4c77-a873-7c586f445773 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.848274] env[62521]: DEBUG nova.compute.provider_tree [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.042061] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318681, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103766} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.042453] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.043577] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80be346e-bc5e-4950-b668-bb139ae8df5b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.065902] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] e2104c11-7713-4542-83da-bc4d2534113a/e2104c11-7713-4542-83da-bc4d2534113a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.066352] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b93b812-643f-40f4-a002-fdf1c44c02b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.085907] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 891.085907] env[62521]: value = "task-1318682" [ 891.085907] env[62521]: _type = "Task" [ 891.085907] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.096507] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318682, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.197792] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318679, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.286777] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318680, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.352765] env[62521]: DEBUG nova.scheduler.client.report [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.596009] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318682, 'name': ReconfigVM_Task, 'duration_secs': 0.403455} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.596373] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Reconfigured VM instance instance-0000004a to attach disk [datastore1] e2104c11-7713-4542-83da-bc4d2534113a/e2104c11-7713-4542-83da-bc4d2534113a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.597015] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f164af20-929f-43e8-bed4-87399d98fdf2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.604130] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 891.604130] env[62521]: value = "task-1318683" [ 891.604130] env[62521]: _type = "Task" [ 891.604130] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.612547] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318683, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.705447] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318679, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.787152] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318680, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.860120] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.860701] env[62521]: DEBUG nova.compute.manager [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.863419] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 15.132s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.115185] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318683, 'name': Rename_Task, 'duration_secs': 0.198802} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.115266] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 892.115545] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be8739cf-679f-4e99-ae85-808f22b2c99c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.122354] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 892.122354] env[62521]: value = "task-1318684" [ 892.122354] env[62521]: _type = "Task" [ 892.122354] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.130465] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318684, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.171768] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "17876bcc-3a7b-47b0-afd9-1215e24864ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.171768] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "17876bcc-3a7b-47b0-afd9-1215e24864ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.198396] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318679, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.46641} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.198665] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9f64b520-b041-469a-8080-6b597c951678/9f64b520-b041-469a-8080-6b597c951678.vmdk to [datastore2] 1d1d2235-748a-4f27-a915-b69f9110257b/1d1d2235-748a-4f27-a915-b69f9110257b.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.199456] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad714e99-8b45-4d98-8220-a07f1315503a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.222689] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 1d1d2235-748a-4f27-a915-b69f9110257b/1d1d2235-748a-4f27-a915-b69f9110257b.vmdk or device None with type streamOptimized {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.223337] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa43f571-a6b6-4b34-b9f9-2593deedcfc4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.243468] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 892.243468] env[62521]: value = "task-1318685" [ 892.243468] env[62521]: _type = "Task" [ 892.243468] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.251895] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318685, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.287381] env[62521]: DEBUG oslo_vmware.api [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318680, 'name': PowerOnVM_Task, 'duration_secs': 2.241749} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.288313] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.288313] env[62521]: INFO nova.compute.manager [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Took 10.80 seconds to spawn the instance on the hypervisor. [ 892.288313] env[62521]: DEBUG nova.compute.manager [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.288879] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f31bec-c6df-403e-9e8c-41f3eb0cc4ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.366795] env[62521]: DEBUG nova.compute.utils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.368232] env[62521]: DEBUG nova.compute.manager [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.368436] env[62521]: DEBUG nova.network.neutron [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 892.372281] env[62521]: INFO nova.compute.claims [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 892.422461] env[62521]: DEBUG nova.policy [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb1d6b613ddc4c8fa4c3bede87ffb00c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4facd01ddac841298d07810781c61e05', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.633058] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318684, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.674343] env[62521]: DEBUG nova.compute.manager [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 892.684416] env[62521]: DEBUG nova.network.neutron [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Successfully created port: c6583bbe-d9ae-4317-8a19-9163ed23ce1d {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.755235] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318685, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.807165] env[62521]: INFO nova.compute.manager [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Took 31.53 seconds to build instance. [ 892.877108] env[62521]: DEBUG nova.compute.manager [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.881614] env[62521]: INFO nova.compute.resource_tracker [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating resource usage from migration 225a66f1-5d99-45a5-a517-4041e11d90f1 [ 893.135201] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318684, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.149858] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75715a9-69b3-443a-8e55-e44f80eaf6c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.157411] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de9a99f-7d6f-4a05-8f47-de3480f676e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.196853] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc73d6df-5a29-4b13-a989-e4ce8366ef76 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.206979] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a1c570d-bcfb-43e1-94e5-53b6b289e44e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.212660] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.221187] env[62521]: DEBUG nova.compute.provider_tree [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.253864] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318685, 'name': ReconfigVM_Task, 'duration_secs': 0.772589} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.254720] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 1d1d2235-748a-4f27-a915-b69f9110257b/1d1d2235-748a-4f27-a915-b69f9110257b.vmdk or device None with type streamOptimized {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.254900] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95cf32cf-3cf1-47ef-a44c-93d79c7181a9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.261573] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 893.261573] env[62521]: value = "task-1318686" [ 893.261573] env[62521]: _type = "Task" [ 893.261573] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.269892] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318686, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.308756] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1b18f255-2444-45a8-a85b-d30330675a96 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "3447712d-daa6-4329-8015-6474cca1a5e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.041s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.454395] env[62521]: DEBUG nova.network.neutron [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Successfully created port: be96c060-332a-43b8-be9e-e62a6dcc3a9f {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.647126] env[62521]: DEBUG oslo_vmware.api [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318684, 'name': PowerOnVM_Task, 'duration_secs': 1.357889} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.650803] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 893.650803] env[62521]: INFO nova.compute.manager [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Took 6.35 seconds to spawn the instance on the hypervisor. [ 893.650803] env[62521]: DEBUG nova.compute.manager [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.650803] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83ad781-190f-4bb9-b41c-b1b2dce70d5a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.725447] env[62521]: DEBUG nova.scheduler.client.report [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.772834] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318686, 'name': Rename_Task, 'duration_secs': 0.14896} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.773125] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.773383] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0805d6ca-3587-4a0f-ab5f-0a88bcc1ff11 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.780555] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 893.780555] env[62521]: value = "task-1318687" [ 893.780555] env[62521]: _type = "Task" [ 893.780555] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.788528] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318687, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.893538] env[62521]: DEBUG nova.compute.manager [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.917851] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.918118] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.918278] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.918464] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.918609] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.918758] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.918967] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.919173] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.919581] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.919581] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.919693] env[62521]: DEBUG nova.virt.hardware [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.920562] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe3df90-7f0b-4dc5-a67e-3e79c505e1e9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.930281] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ceee6e-415a-414b-a45d-dd0abb522c49 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.172143] env[62521]: INFO nova.compute.manager [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Took 29.00 seconds to build instance. [ 894.230044] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.366s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.230318] env[62521]: INFO nova.compute.manager [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Migrating [ 894.237164] env[62521]: DEBUG oslo_concurrency.lockutils [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.494s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.237351] env[62521]: DEBUG nova.objects.instance [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lazy-loading 'resources' on Instance uuid acfd78aa-a51e-4a1c-b01a-15854fe0635c {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.290477] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318687, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.555420] env[62521]: INFO nova.compute.manager [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Rebuilding instance [ 894.596886] env[62521]: DEBUG nova.compute.manager [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.597959] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f8e0ce-19c1-4f9e-9162-5946bcaa84d9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.674582] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8d0c22c8-1d9f-430b-af73-d264796921f0 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "e2104c11-7713-4542-83da-bc4d2534113a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.516s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.746626] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.746842] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.747018] env[62521]: DEBUG nova.network.neutron [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.794930] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318687, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.983010] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6f312c-45ac-419e-8c8a-c2f38eca5d7c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.988976] env[62521]: DEBUG nova.compute.manager [req-9961d636-702b-48b4-aba6-b965cd2931e1 req-9172e5a9-6d61-4e45-84f0-0f9d73a6c89b service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Received event network-vif-plugged-c6583bbe-d9ae-4317-8a19-9163ed23ce1d {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.989273] env[62521]: DEBUG oslo_concurrency.lockutils [req-9961d636-702b-48b4-aba6-b965cd2931e1 req-9172e5a9-6d61-4e45-84f0-0f9d73a6c89b service nova] Acquiring lock "6c4c36b7-b242-4578-9c23-d2529f308cb1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.989517] env[62521]: DEBUG oslo_concurrency.lockutils [req-9961d636-702b-48b4-aba6-b965cd2931e1 req-9172e5a9-6d61-4e45-84f0-0f9d73a6c89b service nova] Lock "6c4c36b7-b242-4578-9c23-d2529f308cb1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.989690] env[62521]: DEBUG oslo_concurrency.lockutils [req-9961d636-702b-48b4-aba6-b965cd2931e1 req-9172e5a9-6d61-4e45-84f0-0f9d73a6c89b service nova] Lock "6c4c36b7-b242-4578-9c23-d2529f308cb1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.989862] env[62521]: DEBUG nova.compute.manager [req-9961d636-702b-48b4-aba6-b965cd2931e1 req-9172e5a9-6d61-4e45-84f0-0f9d73a6c89b service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] No waiting events found dispatching network-vif-plugged-c6583bbe-d9ae-4317-8a19-9163ed23ce1d {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.990043] env[62521]: WARNING nova.compute.manager [req-9961d636-702b-48b4-aba6-b965cd2931e1 req-9172e5a9-6d61-4e45-84f0-0f9d73a6c89b service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Received unexpected event network-vif-plugged-c6583bbe-d9ae-4317-8a19-9163ed23ce1d for instance with vm_state building and task_state spawning. [ 894.995655] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1836050a-2e38-4aaf-9822-477f9975e682 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.034558] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798daf1d-efcf-4b2b-b3e6-2f3deede9572 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.038477] env[62521]: DEBUG nova.compute.manager [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.039493] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f132e46-4f56-498e-b9c6-180704eb30a5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.050380] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32926d62-3ccb-41e2-a66f-784dafb111a6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.067787] env[62521]: DEBUG nova.compute.provider_tree [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.083926] env[62521]: DEBUG nova.network.neutron [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Successfully updated port: c6583bbe-d9ae-4317-8a19-9163ed23ce1d {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.110473] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.110774] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d868ad1-b231-4824-952a-6f0826aeb0cb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.117355] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 895.117355] env[62521]: value = "task-1318688" [ 895.117355] env[62521]: _type = "Task" [ 895.117355] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.125479] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318688, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.292510] env[62521]: DEBUG oslo_vmware.api [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318687, 'name': PowerOnVM_Task, 'duration_secs': 1.312104} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.292833] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.293031] env[62521]: INFO nova.compute.manager [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Took 16.17 seconds to spawn the instance on the hypervisor. [ 895.293198] env[62521]: DEBUG nova.compute.manager [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.293983] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3908af0-458c-4d0b-b49f-2ab117efb41d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.561574] env[62521]: INFO nova.compute.manager [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] instance snapshotting [ 895.564774] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4521e4c1-3da3-4248-b808-f69ddf0e4c1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.570805] env[62521]: DEBUG nova.scheduler.client.report [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.596763] env[62521]: DEBUG oslo_concurrency.lockutils [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.360s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.599274] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ace5257-4f39-4517-921e-9e6dddddf664 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.602153] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.466s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.603914] env[62521]: INFO nova.compute.claims [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.627035] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318688, 'name': PowerOffVM_Task, 'duration_secs': 0.249453} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.628580] env[62521]: INFO nova.scheduler.client.report [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleted allocations for instance acfd78aa-a51e-4a1c-b01a-15854fe0635c [ 895.629847] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.629847] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 895.630481] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ccbade-5daf-47ab-997f-e6cb3218c42f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.640715] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.641030] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca08f7ab-9ae9-4820-9ae8-09b1b4afbbc0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.651558] env[62521]: DEBUG nova.network.neutron [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance_info_cache with network_info: [{"id": "1643da59-fa10-4645-994b-0c557c3a6e10", "address": "fa:16:3e:11:0f:18", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1643da59-fa", "ovs_interfaceid": "1643da59-fa10-4645-994b-0c557c3a6e10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.664698] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.665047] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.665292] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Deleting the datastore file [datastore1] e2104c11-7713-4542-83da-bc4d2534113a {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.666596] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a228ecf-18e8-4570-899d-5384de04c54f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.673514] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 895.673514] env[62521]: value = "task-1318690" [ 895.673514] env[62521]: _type = "Task" [ 895.673514] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.685201] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.817021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.817021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.817021] env[62521]: INFO nova.compute.manager [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Shelving [ 895.820508] env[62521]: INFO nova.compute.manager [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Took 34.61 seconds to build instance. [ 896.117065] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 896.117065] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-04dd9eb9-34dd-4415-8afc-e8a274fcc98b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.125389] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 896.125389] env[62521]: value = "task-1318691" [ 896.125389] env[62521]: _type = "Task" [ 896.125389] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.135473] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318691, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.143614] env[62521]: DEBUG oslo_concurrency.lockutils [None req-33131251-2c50-411f-a66f-ddef253d4ce2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "acfd78aa-a51e-4a1c-b01a-15854fe0635c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.358s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.154869] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.185345] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217948} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.185345] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.185345] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 896.185345] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 896.322542] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5ff34efb-2f6c-4279-89ae-989cd55719fa tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "1d1d2235-748a-4f27-a915-b69f9110257b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.124s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.325472] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.325740] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37835011-5deb-45af-97be-280e22d54a0c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.332814] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 896.332814] env[62521]: value = "task-1318692" [ 896.332814] env[62521]: _type = "Task" [ 896.332814] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.341387] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.581859] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "1d1d2235-748a-4f27-a915-b69f9110257b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.582285] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "1d1d2235-748a-4f27-a915-b69f9110257b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.582607] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "1d1d2235-748a-4f27-a915-b69f9110257b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.582913] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "1d1d2235-748a-4f27-a915-b69f9110257b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.583199] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "1d1d2235-748a-4f27-a915-b69f9110257b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.586060] env[62521]: INFO nova.compute.manager [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Terminating instance [ 896.588452] env[62521]: DEBUG nova.compute.manager [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 896.588719] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 896.589866] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c62dec3d-113b-4487-ae34-5a5e96b6fb1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.599860] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.600189] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d7c10b6-f466-4bd4-916a-e3c911573900 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.606365] env[62521]: DEBUG oslo_vmware.api [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 896.606365] env[62521]: value = "task-1318693" [ 896.606365] env[62521]: _type = "Task" [ 896.606365] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.616875] env[62521]: DEBUG oslo_vmware.api [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318693, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.634812] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318691, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.842290] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318692, 'name': PowerOffVM_Task, 'duration_secs': 0.224668} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.844713] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 896.846023] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d94a3f8-89a5-4852-8280-ab461bf89c37 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.865175] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3376d0bf-b7c1-4e03-851d-d9e181496545 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.868325] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d489e0-3a2f-42ef-ba69-c1e8490e9d49 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.882463] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b98018-73b0-4031-8ce9-2cbb073dc1f2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.914780] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a32131-1bf8-402f-8366-07d7dae368b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.924033] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f44e4a-2856-4979-9a9c-8e5722cdb340 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.940869] env[62521]: DEBUG nova.compute.provider_tree [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.067175] env[62521]: DEBUG nova.compute.manager [req-549890df-ab69-41c6-ae7a-a688062f0303 req-b02d6478-49de-44e5-b8cb-4042cd316767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Received event network-changed-c6583bbe-d9ae-4317-8a19-9163ed23ce1d {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.067438] env[62521]: DEBUG nova.compute.manager [req-549890df-ab69-41c6-ae7a-a688062f0303 req-b02d6478-49de-44e5-b8cb-4042cd316767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Refreshing instance network info cache due to event network-changed-c6583bbe-d9ae-4317-8a19-9163ed23ce1d. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 897.067721] env[62521]: DEBUG oslo_concurrency.lockutils [req-549890df-ab69-41c6-ae7a-a688062f0303 req-b02d6478-49de-44e5-b8cb-4042cd316767 service nova] Acquiring lock "refresh_cache-6c4c36b7-b242-4578-9c23-d2529f308cb1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.067942] env[62521]: DEBUG oslo_concurrency.lockutils [req-549890df-ab69-41c6-ae7a-a688062f0303 req-b02d6478-49de-44e5-b8cb-4042cd316767 service nova] Acquired lock "refresh_cache-6c4c36b7-b242-4578-9c23-d2529f308cb1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.068140] env[62521]: DEBUG nova.network.neutron [req-549890df-ab69-41c6-ae7a-a688062f0303 req-b02d6478-49de-44e5-b8cb-4042cd316767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Refreshing network info cache for port c6583bbe-d9ae-4317-8a19-9163ed23ce1d {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.118104] env[62521]: DEBUG oslo_vmware.api [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318693, 'name': PowerOffVM_Task, 'duration_secs': 0.156636} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.118430] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.118707] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.118978] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3915486c-67b8-4117-bad5-2ccac5f91b94 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.135408] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318691, 'name': CreateSnapshot_Task, 'duration_secs': 0.609576} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.135686] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 897.136544] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d277ca0f-a200-4c63-ae57-8bf6672a44b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.220529] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.220821] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.221012] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleting the datastore file [datastore2] 1d1d2235-748a-4f27-a915-b69f9110257b {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.221289] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccd82a5d-9ead-4139-ad29-1864ffa2f81e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.227610] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.227838] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.228007] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.228207] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.228364] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.228510] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.228718] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.228880] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.229060] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.229232] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.229410] env[62521]: DEBUG nova.virt.hardware [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.230211] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4dcf87-6518-47e1-8a21-b25b05bbdff1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.233613] env[62521]: DEBUG oslo_vmware.api [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 897.233613] env[62521]: value = "task-1318695" [ 897.233613] env[62521]: _type = "Task" [ 897.233613] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.240869] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0e0bad-ea4d-4536-aee9-2798de1c26c3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.248733] env[62521]: DEBUG oslo_vmware.api [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318695, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.258880] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Instance VIF info [] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.264890] env[62521]: DEBUG oslo.service.loopingcall [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.265301] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.265600] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-413464e5-5887-4ae9-9288-129f90be0a1d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.283539] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.283539] env[62521]: value = "task-1318696" [ 897.283539] env[62521]: _type = "Task" [ 897.283539] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.292015] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318696, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.387710] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 897.388112] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bd0aa98f-d382-4b6b-b1e5-0eeaeda81ce1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.400262] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 897.400262] env[62521]: value = "task-1318697" [ 897.400262] env[62521]: _type = "Task" [ 897.400262] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.410137] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318697, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.429874] env[62521]: DEBUG nova.network.neutron [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Successfully updated port: be96c060-332a-43b8-be9e-e62a6dcc3a9f {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 897.444082] env[62521]: DEBUG nova.scheduler.client.report [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.606253] env[62521]: DEBUG nova.network.neutron [req-549890df-ab69-41c6-ae7a-a688062f0303 req-b02d6478-49de-44e5-b8cb-4042cd316767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 897.654240] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 897.656736] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3493fe43-850e-41de-9bbd-09d2da402ee0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.665397] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 897.665397] env[62521]: value = "task-1318698" [ 897.665397] env[62521]: _type = "Task" [ 897.665397] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.671579] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736d7bf4-765c-4ec5-ac9e-81b428a501e3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.676910] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318698, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.691664] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance '3805d897-fcb5-43ed-b502-7fad71124e74' progress to 0 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 897.735915] env[62521]: DEBUG nova.network.neutron [req-549890df-ab69-41c6-ae7a-a688062f0303 req-b02d6478-49de-44e5-b8cb-4042cd316767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.746604] env[62521]: DEBUG oslo_vmware.api [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318695, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.233281} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.746884] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.747086] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 897.747276] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 897.747448] env[62521]: INFO nova.compute.manager [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Took 1.16 seconds to destroy the instance on the hypervisor. [ 897.747690] env[62521]: DEBUG oslo.service.loopingcall [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.748123] env[62521]: DEBUG nova.compute.manager [-] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 897.748226] env[62521]: DEBUG nova.network.neutron [-] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 897.799201] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318696, 'name': CreateVM_Task, 'duration_secs': 0.298695} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.799201] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 897.799201] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.799201] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.799201] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.799201] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd8d2081-c990-42d1-8342-b85a780b8eb5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.801856] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 897.801856] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fb7688-a7ca-6cf8-69d2-84b73c747172" [ 897.801856] env[62521]: _type = "Task" [ 897.801856] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.810975] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fb7688-a7ca-6cf8-69d2-84b73c747172, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.910020] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318697, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.933080] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "refresh_cache-6c4c36b7-b242-4578-9c23-d2529f308cb1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.949167] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.949729] env[62521]: DEBUG nova.compute.manager [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.952587] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.677s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.954154] env[62521]: INFO nova.compute.claims [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.176669] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318698, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.199193] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 898.199193] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bdeabb64-870f-42b8-b37b-42df6c0fb2ee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.205120] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 898.205120] env[62521]: value = "task-1318699" [ 898.205120] env[62521]: _type = "Task" [ 898.205120] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.215240] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.241801] env[62521]: DEBUG oslo_concurrency.lockutils [req-549890df-ab69-41c6-ae7a-a688062f0303 req-b02d6478-49de-44e5-b8cb-4042cd316767 service nova] Releasing lock "refresh_cache-6c4c36b7-b242-4578-9c23-d2529f308cb1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.242248] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquired lock "refresh_cache-6c4c36b7-b242-4578-9c23-d2529f308cb1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.242428] env[62521]: DEBUG nova.network.neutron [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 898.314870] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fb7688-a7ca-6cf8-69d2-84b73c747172, 'name': SearchDatastore_Task, 'duration_secs': 0.012381} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.315237] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.315482] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.315735] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.315915] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.316103] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.316434] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd6d46d2-77e7-4c04-82ab-a09f10706943 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.324450] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.324653] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.325472] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd18e50a-885b-4be3-957d-d9f25b91625b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.331153] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 898.331153] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52dd78b9-29dd-438a-6988-5b466ea3a06c" [ 898.331153] env[62521]: _type = "Task" [ 898.331153] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.339059] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52dd78b9-29dd-438a-6988-5b466ea3a06c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.410739] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318697, 'name': CreateSnapshot_Task, 'duration_secs': 0.711416} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.411571] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 898.412089] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ef70c9-072b-466e-91d3-6194e4c89a17 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.460160] env[62521]: DEBUG nova.compute.utils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.463040] env[62521]: DEBUG nova.compute.manager [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 898.463240] env[62521]: DEBUG nova.network.neutron [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 898.511277] env[62521]: DEBUG nova.policy [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfab7121825d4e049370b526f90eb499', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11829634203c4967bbcff90eb904097c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.676285] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318698, 'name': CloneVM_Task} progress is 95%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.714325] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318699, 'name': PowerOffVM_Task, 'duration_secs': 0.163935} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.714600] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 898.714784] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance '3805d897-fcb5-43ed-b502-7fad71124e74' progress to 17 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 898.744431] env[62521]: DEBUG nova.network.neutron [-] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.769097] env[62521]: DEBUG nova.network.neutron [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Successfully created port: 9e954d34-4424-459c-a022-fc039668c616 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.798092] env[62521]: DEBUG nova.network.neutron [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 898.841640] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52dd78b9-29dd-438a-6988-5b466ea3a06c, 'name': SearchDatastore_Task, 'duration_secs': 0.00911} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.842268] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e486de44-b096-4a3a-9d6c-9e1c9fac4917 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.847644] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 898.847644] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521ba079-79e1-2b4b-280a-c0816229be38" [ 898.847644] env[62521]: _type = "Task" [ 898.847644] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.855392] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521ba079-79e1-2b4b-280a-c0816229be38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.934138] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 898.934565] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b07e29f3-d147-4079-9352-f87f358a5e72 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.943567] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 898.943567] env[62521]: value = "task-1318700" [ 898.943567] env[62521]: _type = "Task" [ 898.943567] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.951726] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318700, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.967509] env[62521]: DEBUG nova.compute.manager [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 899.183069] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318698, 'name': CloneVM_Task, 'duration_secs': 1.497693} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.183582] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Created linked-clone VM from snapshot [ 899.184163] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7ba14d-e49a-411f-88b1-513d4a567279 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.198696] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Uploading image b533ce3c-5cca-413e-961f-fa0513cb3e05 {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 899.215818] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 899.216156] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-03d39289-fbd0-4b82-bc60-14ee610e2016 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.222686] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.222948] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.223317] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.223543] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.223703] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.223861] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.224086] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.224256] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.224434] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.224602] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.224783] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.235044] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34c0fc63-ad0a-437c-b88c-e0f3cbfecb96 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.245354] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 899.245354] env[62521]: value = "task-1318701" [ 899.245354] env[62521]: _type = "Task" [ 899.245354] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.249493] env[62521]: INFO nova.compute.manager [-] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Took 1.50 seconds to deallocate network for instance. [ 899.256319] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 899.256319] env[62521]: value = "task-1318702" [ 899.256319] env[62521]: _type = "Task" [ 899.256319] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.258292] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0dbc33-3d56-43c3-9a1a-3e72ce821863 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.267710] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318701, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.275924] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9c736d-42ca-4cac-9838-c6e298615276 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.279818] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318702, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.316095] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e2d18e-d9e6-4c84-8ce7-616da4b9b374 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.320694] env[62521]: DEBUG nova.compute.manager [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Received event network-vif-plugged-be96c060-332a-43b8-be9e-e62a6dcc3a9f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.321291] env[62521]: DEBUG oslo_concurrency.lockutils [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] Acquiring lock "6c4c36b7-b242-4578-9c23-d2529f308cb1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.321416] env[62521]: DEBUG oslo_concurrency.lockutils [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] Lock "6c4c36b7-b242-4578-9c23-d2529f308cb1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.321726] env[62521]: DEBUG oslo_concurrency.lockutils [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] Lock "6c4c36b7-b242-4578-9c23-d2529f308cb1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.323136] env[62521]: DEBUG nova.compute.manager [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] No waiting events found dispatching network-vif-plugged-be96c060-332a-43b8-be9e-e62a6dcc3a9f {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.323136] env[62521]: WARNING nova.compute.manager [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Received unexpected event network-vif-plugged-be96c060-332a-43b8-be9e-e62a6dcc3a9f for instance with vm_state building and task_state spawning. [ 899.323136] env[62521]: DEBUG nova.compute.manager [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Received event network-changed-be96c060-332a-43b8-be9e-e62a6dcc3a9f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.323429] env[62521]: DEBUG nova.compute.manager [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Refreshing instance network info cache due to event network-changed-be96c060-332a-43b8-be9e-e62a6dcc3a9f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 899.323469] env[62521]: DEBUG oslo_concurrency.lockutils [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] Acquiring lock "refresh_cache-6c4c36b7-b242-4578-9c23-d2529f308cb1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.332962] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1b317e-9b24-4d5b-854f-18e075e9d14c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.357963] env[62521]: DEBUG nova.compute.provider_tree [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.375970] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521ba079-79e1-2b4b-280a-c0816229be38, 'name': SearchDatastore_Task, 'duration_secs': 0.012464} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.376314] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.376714] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] e2104c11-7713-4542-83da-bc4d2534113a/e2104c11-7713-4542-83da-bc4d2534113a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 899.377417] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5eed16e3-0ef0-4bf9-9d02-7b2483ec708a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.386992] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 899.386992] env[62521]: value = "task-1318703" [ 899.386992] env[62521]: _type = "Task" [ 899.386992] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.399768] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.456738] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318700, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.756712] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.757050] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318701, 'name': Destroy_Task, 'duration_secs': 0.391365} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.757326] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Destroyed the VM [ 899.757570] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 899.757885] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-298212c9-dd7b-4766-a3cd-a940821f302e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.768855] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318702, 'name': ReconfigVM_Task, 'duration_secs': 0.325051} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.770338] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance '3805d897-fcb5-43ed-b502-7fad71124e74' progress to 33 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 899.774294] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 899.774294] env[62521]: value = "task-1318704" [ 899.774294] env[62521]: _type = "Task" [ 899.774294] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.775460] env[62521]: DEBUG nova.network.neutron [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Updating instance_info_cache with network_info: [{"id": "c6583bbe-d9ae-4317-8a19-9163ed23ce1d", "address": "fa:16:3e:39:a8:9e", "network": {"id": "bdf4afe5-9e72-407e-863f-affd552d396d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1173988657", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.151", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47499d09-8010-4d02-ac96-4f057c104692", "external-id": "nsx-vlan-transportzone-14", "segmentation_id": 14, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6583bbe-d9", "ovs_interfaceid": "c6583bbe-d9ae-4317-8a19-9163ed23ce1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "be96c060-332a-43b8-be9e-e62a6dcc3a9f", "address": "fa:16:3e:57:f4:8a", "network": {"id": "3b7e8a37-2525-4c2e-bf2d-532ad22f8fff", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-379131705", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.192", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe96c060-33", "ovs_interfaceid": "be96c060-332a-43b8-be9e-e62a6dcc3a9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.786057] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318704, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.867154] env[62521]: DEBUG nova.scheduler.client.report [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.897936] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318703, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.956111] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318700, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.982716] env[62521]: DEBUG nova.compute.manager [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 900.010941] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.011313] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.011511] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.011747] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.011935] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.012116] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.012398] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.012578] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.012783] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.013023] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.013243] env[62521]: DEBUG nova.virt.hardware [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.014364] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9fe4053-38d1-4de5-ba14-8cc387d81b77 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.023730] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6795c618-31cc-4db7-a5ff-6081481ae0f8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.279050] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Releasing lock "refresh_cache-6c4c36b7-b242-4578-9c23-d2529f308cb1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.279457] env[62521]: DEBUG nova.compute.manager [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Instance network_info: |[{"id": "c6583bbe-d9ae-4317-8a19-9163ed23ce1d", "address": "fa:16:3e:39:a8:9e", "network": {"id": "bdf4afe5-9e72-407e-863f-affd552d396d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1173988657", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.151", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47499d09-8010-4d02-ac96-4f057c104692", "external-id": "nsx-vlan-transportzone-14", "segmentation_id": 14, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6583bbe-d9", "ovs_interfaceid": "c6583bbe-d9ae-4317-8a19-9163ed23ce1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "be96c060-332a-43b8-be9e-e62a6dcc3a9f", "address": "fa:16:3e:57:f4:8a", "network": {"id": "3b7e8a37-2525-4c2e-bf2d-532ad22f8fff", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-379131705", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.192", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe96c060-33", "ovs_interfaceid": "be96c060-332a-43b8-be9e-e62a6dcc3a9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 900.281689] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.281918] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.282095] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.282290] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.282449] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.282610] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.282820] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.283013] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.283204] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.283381] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.283556] env[62521]: DEBUG nova.virt.hardware [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.289167] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Reconfiguring VM instance instance-00000045 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 900.289506] env[62521]: DEBUG oslo_concurrency.lockutils [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] Acquired lock "refresh_cache-6c4c36b7-b242-4578-9c23-d2529f308cb1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.289693] env[62521]: DEBUG nova.network.neutron [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Refreshing network info cache for port be96c060-332a-43b8-be9e-e62a6dcc3a9f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.290989] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:a8:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '47499d09-8010-4d02-ac96-4f057c104692', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6583bbe-d9ae-4317-8a19-9163ed23ce1d', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:f4:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4aa1eda7-48b9-4fa2-af0b-94c718313af2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be96c060-332a-43b8-be9e-e62a6dcc3a9f', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.300870] env[62521]: DEBUG oslo.service.loopingcall [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.304615] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6e5ed77-21ed-4670-bcd0-0ed2f44ad633 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.318542] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.318795] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7feea6f2-4029-4451-a39f-d92fcce72906 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.341691] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318704, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.343372] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.343372] env[62521]: value = "task-1318706" [ 900.343372] env[62521]: _type = "Task" [ 900.343372] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.343664] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 900.343664] env[62521]: value = "task-1318705" [ 900.343664] env[62521]: _type = "Task" [ 900.343664] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.354334] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318706, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.357066] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.373247] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.373945] env[62521]: DEBUG nova.compute.manager [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 900.376966] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.589s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.378517] env[62521]: INFO nova.compute.claims [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.399174] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.973608} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.399444] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] e2104c11-7713-4542-83da-bc4d2534113a/e2104c11-7713-4542-83da-bc4d2534113a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 900.399665] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.399916] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5954665d-cc8c-467b-b78b-c56ff8be0253 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.406442] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 900.406442] env[62521]: value = "task-1318707" [ 900.406442] env[62521]: _type = "Task" [ 900.406442] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.416668] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318707, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.455092] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318700, 'name': CloneVM_Task} progress is 95%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.789802] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318704, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.858087] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318705, 'name': ReconfigVM_Task, 'duration_secs': 0.197722} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.861092] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Reconfigured VM instance instance-00000045 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 900.861345] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318706, 'name': CreateVM_Task, 'duration_secs': 0.457679} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.862016] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384e55e0-78db-4185-ade9-eeb080464d34 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.864393] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 900.865106] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.865301] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.865682] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.866116] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99a91768-a53c-473a-908c-c6152d2d8d77 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.889084] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 3805d897-fcb5-43ed-b502-7fad71124e74/3805d897-fcb5-43ed-b502-7fad71124e74.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.897854] env[62521]: DEBUG nova.compute.utils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 900.899490] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0af440c-d85e-4852-aafc-7f7e6b8f3b35 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.915548] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 900.915548] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521b131b-888f-86f6-cc97-c5842b1fc5f0" [ 900.915548] env[62521]: _type = "Task" [ 900.915548] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.919784] env[62521]: DEBUG nova.compute.manager [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 900.924462] env[62521]: DEBUG nova.compute.manager [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 900.924652] env[62521]: DEBUG nova.network.neutron [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 900.938900] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 900.938900] env[62521]: value = "task-1318708" [ 900.938900] env[62521]: _type = "Task" [ 900.938900] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.945642] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318707, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.145325} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.945965] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521b131b-888f-86f6-cc97-c5842b1fc5f0, 'name': SearchDatastore_Task, 'duration_secs': 0.028197} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.949460] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 900.952527] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.952792] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.953080] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.953210] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.953396] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.957275] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2bc6d23-df3d-4421-bc7c-5e9b559c40f2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.959788] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2317a5b5-8769-4173-b6c1-002394d2c220 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.968226] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318708, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.988927] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] e2104c11-7713-4542-83da-bc4d2534113a/e2104c11-7713-4542-83da-bc4d2534113a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.988927] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318700, 'name': CloneVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.991031] env[62521]: DEBUG nova.policy [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba9b0b82d53e4502bef18a9f031fcc9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5780fc8ecfd64161bc2ea0b3a092b4d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 900.992699] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c97a29eb-0b66-4c41-a2de-26f525715aa8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.006803] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.007039] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.008079] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c52845c-778e-4a20-95f3-da646a2b548d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.014545] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 901.014545] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526f22ae-28f3-a4c3-99a0-bf04ae2019c4" [ 901.014545] env[62521]: _type = "Task" [ 901.014545] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.019558] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 901.019558] env[62521]: value = "task-1318709" [ 901.019558] env[62521]: _type = "Task" [ 901.019558] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.028210] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526f22ae-28f3-a4c3-99a0-bf04ae2019c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.033328] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318709, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.248970] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdac5ef-f52f-4b13-bb43-c0ee70d0e4d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.257225] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b8b9224-73e9-43a0-94d3-baafd41d027a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.302100] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3882f4bc-49fd-44f8-972d-12f4a5d49322 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.334787] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438cefbc-3cd8-4051-af52-e9d876bf1042 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.334787] env[62521]: DEBUG oslo_vmware.api [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318704, 'name': RemoveSnapshot_Task, 'duration_secs': 1.425711} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.334787] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 901.334787] env[62521]: DEBUG nova.compute.provider_tree [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.339430] env[62521]: DEBUG nova.network.neutron [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Successfully created port: 2c0a3f29-63a4-44ff-bbf5-40908d17dc42 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 901.368855] env[62521]: DEBUG nova.compute.manager [req-d2566a4f-42ae-4edc-98f4-9ea3758e0d95 req-f6be99e4-622a-4057-98dd-9aa41b6ca323 service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Received event network-vif-plugged-9e954d34-4424-459c-a022-fc039668c616 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.369099] env[62521]: DEBUG oslo_concurrency.lockutils [req-d2566a4f-42ae-4edc-98f4-9ea3758e0d95 req-f6be99e4-622a-4057-98dd-9aa41b6ca323 service nova] Acquiring lock "11e61e72-a311-4c43-bf53-df82ad9d70f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.369327] env[62521]: DEBUG oslo_concurrency.lockutils [req-d2566a4f-42ae-4edc-98f4-9ea3758e0d95 req-f6be99e4-622a-4057-98dd-9aa41b6ca323 service nova] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.369500] env[62521]: DEBUG oslo_concurrency.lockutils [req-d2566a4f-42ae-4edc-98f4-9ea3758e0d95 req-f6be99e4-622a-4057-98dd-9aa41b6ca323 service nova] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.369853] env[62521]: DEBUG nova.compute.manager [req-d2566a4f-42ae-4edc-98f4-9ea3758e0d95 req-f6be99e4-622a-4057-98dd-9aa41b6ca323 service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] No waiting events found dispatching network-vif-plugged-9e954d34-4424-459c-a022-fc039668c616 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 901.369853] env[62521]: WARNING nova.compute.manager [req-d2566a4f-42ae-4edc-98f4-9ea3758e0d95 req-f6be99e4-622a-4057-98dd-9aa41b6ca323 service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Received unexpected event network-vif-plugged-9e954d34-4424-459c-a022-fc039668c616 for instance with vm_state building and task_state spawning. [ 901.411448] env[62521]: DEBUG nova.network.neutron [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Updated VIF entry in instance network info cache for port be96c060-332a-43b8-be9e-e62a6dcc3a9f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.412066] env[62521]: DEBUG nova.network.neutron [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Updating instance_info_cache with network_info: [{"id": "c6583bbe-d9ae-4317-8a19-9163ed23ce1d", "address": "fa:16:3e:39:a8:9e", "network": {"id": "bdf4afe5-9e72-407e-863f-affd552d396d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1173988657", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.151", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47499d09-8010-4d02-ac96-4f057c104692", "external-id": "nsx-vlan-transportzone-14", "segmentation_id": 14, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6583bbe-d9", "ovs_interfaceid": "c6583bbe-d9ae-4317-8a19-9163ed23ce1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "be96c060-332a-43b8-be9e-e62a6dcc3a9f", "address": "fa:16:3e:57:f4:8a", "network": {"id": "3b7e8a37-2525-4c2e-bf2d-532ad22f8fff", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-379131705", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.192", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4facd01ddac841298d07810781c61e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe96c060-33", "ovs_interfaceid": "be96c060-332a-43b8-be9e-e62a6dcc3a9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.457461] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318700, 'name': CloneVM_Task, 'duration_secs': 2.034958} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.460517] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Created linked-clone VM from snapshot [ 901.460800] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318708, 'name': ReconfigVM_Task, 'duration_secs': 0.334476} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.461558] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d673e05d-68ae-4d18-9154-e65b37d8525a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.463935] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 3805d897-fcb5-43ed-b502-7fad71124e74/3805d897-fcb5-43ed-b502-7fad71124e74.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.464234] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance '3805d897-fcb5-43ed-b502-7fad71124e74' progress to 50 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 901.474109] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Uploading image 720eb7fe-5500-4cf9-a06e-3471ca9179ed {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 901.499236] env[62521]: DEBUG oslo_vmware.rw_handles [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 901.499236] env[62521]: value = "vm-282123" [ 901.499236] env[62521]: _type = "VirtualMachine" [ 901.499236] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 901.499821] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-cc5adc89-5998-4403-96f2-a01e10e42c6a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.506766] env[62521]: DEBUG oslo_vmware.rw_handles [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lease: (returnval){ [ 901.506766] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e07ac-b297-988b-3644-103b7112bc71" [ 901.506766] env[62521]: _type = "HttpNfcLease" [ 901.506766] env[62521]: } obtained for exporting VM: (result){ [ 901.506766] env[62521]: value = "vm-282123" [ 901.506766] env[62521]: _type = "VirtualMachine" [ 901.506766] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 901.507373] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the lease: (returnval){ [ 901.507373] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e07ac-b297-988b-3644-103b7112bc71" [ 901.507373] env[62521]: _type = "HttpNfcLease" [ 901.507373] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 901.513780] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 901.513780] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e07ac-b297-988b-3644-103b7112bc71" [ 901.513780] env[62521]: _type = "HttpNfcLease" [ 901.513780] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 901.524734] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526f22ae-28f3-a4c3-99a0-bf04ae2019c4, 'name': SearchDatastore_Task, 'duration_secs': 0.01245} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.528088] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34a55745-b619-4366-8349-9dc020356c1b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.537014] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 901.537014] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e6b823-c95d-f492-26b1-a035311206e7" [ 901.537014] env[62521]: _type = "Task" [ 901.537014] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.537259] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318709, 'name': ReconfigVM_Task, 'duration_secs': 0.328614} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.539135] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Reconfigured VM instance instance-0000004a to attach disk [datastore2] e2104c11-7713-4542-83da-bc4d2534113a/e2104c11-7713-4542-83da-bc4d2534113a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.540623] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39e0c6bd-8b91-4ac9-918d-249297430733 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.542883] env[62521]: DEBUG nova.network.neutron [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Successfully updated port: 9e954d34-4424-459c-a022-fc039668c616 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.549700] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e6b823-c95d-f492-26b1-a035311206e7, 'name': SearchDatastore_Task, 'duration_secs': 0.00889} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.551019] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.551178] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 6c4c36b7-b242-4578-9c23-d2529f308cb1/6c4c36b7-b242-4578-9c23-d2529f308cb1.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 901.551733] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 901.551733] env[62521]: value = "task-1318711" [ 901.551733] env[62521]: _type = "Task" [ 901.551733] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.551733] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4928a218-bc79-4069-80f3-f1e707ef135f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.563799] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318711, 'name': Rename_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.565138] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 901.565138] env[62521]: value = "task-1318712" [ 901.565138] env[62521]: _type = "Task" [ 901.565138] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.574210] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318712, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.821783] env[62521]: WARNING nova.compute.manager [None req-cdc82ed3-96e7-45ff-83fb-c5f3b25bca8b tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Image not found during snapshot: nova.exception.ImageNotFound: Image b533ce3c-5cca-413e-961f-fa0513cb3e05 could not be found. [ 901.835131] env[62521]: DEBUG nova.scheduler.client.report [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.917251] env[62521]: DEBUG oslo_concurrency.lockutils [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] Releasing lock "refresh_cache-6c4c36b7-b242-4578-9c23-d2529f308cb1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.917251] env[62521]: DEBUG nova.compute.manager [req-b52ace22-bfc5-4ba2-98b0-e2b3033d2f94 req-e06bb513-82c2-4f65-9181-ba8821e29767 service nova] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Received event network-vif-deleted-80906e15-3b2b-48f3-8c94-b2061848d775 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.933697] env[62521]: DEBUG nova.compute.manager [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.957679] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.958019] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.958199] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.958432] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.958593] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.958735] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.958941] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.959134] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.959318] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.959531] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.959711] env[62521]: DEBUG nova.virt.hardware [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.960725] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699b060b-4a12-4b60-894b-04bfcf049ea7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.972160] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8e2bed-43a6-4e35-a1ae-0bd1178f6b5c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.975714] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4669978-8a04-4b1e-b326-c885388af75d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.008663] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8668a2-e59e-4d64-9e6f-b5a9c863b371 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.024654] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.024654] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e07ac-b297-988b-3644-103b7112bc71" [ 902.024654] env[62521]: _type = "HttpNfcLease" [ 902.024654] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 902.739294] env[62521]: DEBUG oslo_vmware.rw_handles [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 902.739294] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e07ac-b297-988b-3644-103b7112bc71" [ 902.739294] env[62521]: _type = "HttpNfcLease" [ 902.739294] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 902.739875] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.739997] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.740160] env[62521]: DEBUG nova.network.neutron [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.743019] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.743482] env[62521]: DEBUG nova.compute.manager [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 902.745854] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "3447712d-daa6-4329-8015-6474cca1a5e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.746081] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "3447712d-daa6-4329-8015-6474cca1a5e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.746427] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "3447712d-daa6-4329-8015-6474cca1a5e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.746628] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "3447712d-daa6-4329-8015-6474cca1a5e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.746843] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "3447712d-daa6-4329-8015-6474cca1a5e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.748397] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance '3805d897-fcb5-43ed-b502-7fad71124e74' progress to 67 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 902.751460] env[62521]: INFO nova.compute.manager [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Terminating instance [ 902.753408] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b464b3-0250-44fd-984f-815ea3bb753c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.762257] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.538s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.764910] env[62521]: DEBUG nova.compute.manager [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 902.766172] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 902.766282] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cb5ce5-c48f-413d-a7ee-394f0627b416 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.774549] env[62521]: DEBUG oslo_vmware.rw_handles [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e4633-3ce5-0d2d-7a92-4d41e45e1ae1/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 902.774740] env[62521]: DEBUG oslo_vmware.rw_handles [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e4633-3ce5-0d2d-7a92-4d41e45e1ae1/disk-0.vmdk for reading. {{(pid=62521) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 902.783456] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318711, 'name': Rename_Task, 'duration_secs': 0.15261} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.783691] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318712, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495964} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.784186] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 902.785232] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 902.785482] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 6c4c36b7-b242-4578-9c23-d2529f308cb1/6c4c36b7-b242-4578-9c23-d2529f308cb1.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 902.785688] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 902.785900] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d4b113a-3ab9-4c4b-b612-6a4accd944a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.839978] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-37b51068-582c-4cdb-9294-11849966124c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.842319] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d3b771e-bd86-474a-afb3-e0b6590968ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.851047] env[62521]: DEBUG nova.compute.manager [req-36cc8376-6c75-473f-8aed-188acf6198ab req-5d86141f-3dea-4fb6-868e-e5371eb4b777 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Received event network-vif-plugged-2c0a3f29-63a4-44ff-bbf5-40908d17dc42 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.851047] env[62521]: DEBUG oslo_concurrency.lockutils [req-36cc8376-6c75-473f-8aed-188acf6198ab req-5d86141f-3dea-4fb6-868e-e5371eb4b777 service nova] Acquiring lock "563ad3b2-e53e-4682-9cde-bd6f709718de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.851047] env[62521]: DEBUG oslo_concurrency.lockutils [req-36cc8376-6c75-473f-8aed-188acf6198ab req-5d86141f-3dea-4fb6-868e-e5371eb4b777 service nova] Lock "563ad3b2-e53e-4682-9cde-bd6f709718de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.851047] env[62521]: DEBUG oslo_concurrency.lockutils [req-36cc8376-6c75-473f-8aed-188acf6198ab req-5d86141f-3dea-4fb6-868e-e5371eb4b777 service nova] Lock "563ad3b2-e53e-4682-9cde-bd6f709718de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.851047] env[62521]: DEBUG nova.compute.manager [req-36cc8376-6c75-473f-8aed-188acf6198ab req-5d86141f-3dea-4fb6-868e-e5371eb4b777 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] No waiting events found dispatching network-vif-plugged-2c0a3f29-63a4-44ff-bbf5-40908d17dc42 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.851047] env[62521]: WARNING nova.compute.manager [req-36cc8376-6c75-473f-8aed-188acf6198ab req-5d86141f-3dea-4fb6-868e-e5371eb4b777 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Received unexpected event network-vif-plugged-2c0a3f29-63a4-44ff-bbf5-40908d17dc42 for instance with vm_state building and task_state spawning. [ 902.853147] env[62521]: DEBUG oslo_vmware.api [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 902.853147] env[62521]: value = "task-1318713" [ 902.853147] env[62521]: _type = "Task" [ 902.853147] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.858825] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 902.858825] env[62521]: value = "task-1318715" [ 902.858825] env[62521]: _type = "Task" [ 902.858825] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.865342] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 902.865342] env[62521]: value = "task-1318714" [ 902.865342] env[62521]: _type = "Task" [ 902.865342] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.865342] env[62521]: DEBUG nova.network.neutron [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Successfully updated port: 2c0a3f29-63a4-44ff-bbf5-40908d17dc42 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.874311] env[62521]: DEBUG oslo_vmware.api [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318713, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.880410] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318715, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.884403] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318714, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.889741] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-84817e16-5606-4751-85ff-a12765bc7d58 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.254331] env[62521]: DEBUG nova.compute.utils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 903.258153] env[62521]: DEBUG nova.compute.manager [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 903.258416] env[62521]: DEBUG nova.network.neutron [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 903.265058] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.265319] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.265513] env[62521]: DEBUG nova.compute.manager [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.274189] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25dd09ed-9883-4b0b-9208-a518de6d6684 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.285968] env[62521]: DEBUG nova.compute.manager [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62521) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 903.287868] env[62521]: DEBUG nova.objects.instance [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'flavor' on Instance uuid 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.292749] env[62521]: DEBUG nova.network.neutron [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 903.343609] env[62521]: DEBUG nova.policy [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76d85009c2c1416b9ea049003e6d96a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c533bd4917b4466e9c5c76a9caf8e807', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 903.363215] env[62521]: DEBUG oslo_vmware.api [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318713, 'name': PowerOffVM_Task, 'duration_secs': 0.217046} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.366785] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 903.367230] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.367720] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd076309-c249-4a17-8b83-79d1296a0d45 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.376589] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "refresh_cache-563ad3b2-e53e-4682-9cde-bd6f709718de" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.379484] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "refresh_cache-563ad3b2-e53e-4682-9cde-bd6f709718de" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.379643] env[62521]: DEBUG nova.network.neutron [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 903.387051] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318715, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07173} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.392508] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 903.397428] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-315fc8f4-84d6-42a5-98c4-e33288fd60d5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.400901] env[62521]: DEBUG oslo_vmware.api [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318714, 'name': PowerOnVM_Task, 'duration_secs': 0.531414} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.402171] env[62521]: DEBUG nova.network.neutron [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Port 1643da59-fa10-4645-994b-0c557c3a6e10 binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 903.403784] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 903.404008] env[62521]: DEBUG nova.compute.manager [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.405432] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0798b985-7581-49aa-8aa6-4cb5f1909562 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.436353] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 6c4c36b7-b242-4578-9c23-d2529f308cb1/6c4c36b7-b242-4578-9c23-d2529f308cb1.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.440622] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce1b5b9b-e7be-4aba-88bf-c5082e9e7d21 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.461793] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.462139] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.462455] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Deleting the datastore file [datastore2] 3447712d-daa6-4329-8015-6474cca1a5e4 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.466274] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa1509dd-5e6a-4a9e-8f61-a3dea1837177 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.470670] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 903.470670] env[62521]: value = "task-1318717" [ 903.470670] env[62521]: _type = "Task" [ 903.470670] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.475770] env[62521]: DEBUG oslo_vmware.api [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 903.475770] env[62521]: value = "task-1318718" [ 903.475770] env[62521]: _type = "Task" [ 903.475770] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.482967] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318717, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.488624] env[62521]: DEBUG oslo_vmware.api [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318718, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.489767] env[62521]: DEBUG nova.network.neutron [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 903.513370] env[62521]: DEBUG nova.network.neutron [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance_info_cache with network_info: [{"id": "9e954d34-4424-459c-a022-fc039668c616", "address": "fa:16:3e:90:66:a0", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e954d34-44", "ovs_interfaceid": "9e954d34-4424-459c-a022-fc039668c616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.592462] env[62521]: DEBUG nova.compute.manager [req-4564dd60-3528-42ab-9059-9215a0f0bb0f req-721e5971-8754-49c4-a33b-d693f5789a0c service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Received event network-changed-9e954d34-4424-459c-a022-fc039668c616 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.592775] env[62521]: DEBUG nova.compute.manager [req-4564dd60-3528-42ab-9059-9215a0f0bb0f req-721e5971-8754-49c4-a33b-d693f5789a0c service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Refreshing instance network info cache due to event network-changed-9e954d34-4424-459c-a022-fc039668c616. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 903.593073] env[62521]: DEBUG oslo_concurrency.lockutils [req-4564dd60-3528-42ab-9059-9215a0f0bb0f req-721e5971-8754-49c4-a33b-d693f5789a0c service nova] Acquiring lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.759205] env[62521]: DEBUG nova.compute.manager [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 903.782285] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Applying migration context for instance 3805d897-fcb5-43ed-b502-7fad71124e74 as it has an incoming, in-progress migration 225a66f1-5d99-45a5-a517-4041e11d90f1. Migration status is migrating {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 903.785316] env[62521]: INFO nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating resource usage from migration 225a66f1-5d99-45a5-a517-4041e11d90f1 [ 903.793511] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.793786] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7dbc21af-bbb8-4264-a009-b62e2fa652dd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.800730] env[62521]: DEBUG oslo_vmware.api [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 903.800730] env[62521]: value = "task-1318719" [ 903.800730] env[62521]: _type = "Task" [ 903.800730] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.811341] env[62521]: DEBUG oslo_vmware.api [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.818936] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.819163] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance af91de53-2ef8-48d1-b8c1-7bd0809f4ae4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.819304] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 4dcd1c6c-8726-42db-997a-e78ba1293310 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.819512] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.819592] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.819677] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.819794] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance d5f6492d-ca1b-4e74-b792-b04d55c33660 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.819905] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance d574f077-b221-4f91-8b54-0915421cb36f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.820051] env[62521]: WARNING nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 1d1d2235-748a-4f27-a915-b69f9110257b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 903.820168] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 3447712d-daa6-4329-8015-6474cca1a5e4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.820280] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance e2104c11-7713-4542-83da-bc4d2534113a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.820389] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 6c4c36b7-b242-4578-9c23-d2529f308cb1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.820502] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Migration 225a66f1-5d99-45a5-a517-4041e11d90f1 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 903.820611] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 3805d897-fcb5-43ed-b502-7fad71124e74 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.820736] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 11e61e72-a311-4c43-bf53-df82ad9d70f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.820847] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 563ad3b2-e53e-4682-9cde-bd6f709718de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.820954] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 611462d2-7f57-44b0-ac36-db32af3d0dd0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.995797] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.003717] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318717, 'name': ReconfigVM_Task, 'duration_secs': 0.332491} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.007435] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 6c4c36b7-b242-4578-9c23-d2529f308cb1/6c4c36b7-b242-4578-9c23-d2529f308cb1.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.008199] env[62521]: DEBUG oslo_vmware.api [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318718, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192144} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.008462] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4a2f4657-d67e-4f87-a1a8-c9361fd6c4a2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.010696] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.010983] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.011199] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.011772] env[62521]: INFO nova.compute.manager [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Took 1.25 seconds to destroy the instance on the hypervisor. [ 904.012219] env[62521]: DEBUG oslo.service.loopingcall [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.012745] env[62521]: DEBUG nova.compute.manager [-] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 904.012969] env[62521]: DEBUG nova.network.neutron [-] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 904.015753] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.016146] env[62521]: DEBUG nova.compute.manager [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Instance network_info: |[{"id": "9e954d34-4424-459c-a022-fc039668c616", "address": "fa:16:3e:90:66:a0", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e954d34-44", "ovs_interfaceid": "9e954d34-4424-459c-a022-fc039668c616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 904.016487] env[62521]: DEBUG oslo_concurrency.lockutils [req-4564dd60-3528-42ab-9059-9215a0f0bb0f req-721e5971-8754-49c4-a33b-d693f5789a0c service nova] Acquired lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.016760] env[62521]: DEBUG nova.network.neutron [req-4564dd60-3528-42ab-9059-9215a0f0bb0f req-721e5971-8754-49c4-a33b-d693f5789a0c service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Refreshing network info cache for port 9e954d34-4424-459c-a022-fc039668c616 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.018248] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:66:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e954d34-4424-459c-a022-fc039668c616', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.026045] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating folder: Project (11829634203c4967bbcff90eb904097c). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 904.032443] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-57ff055f-1cc1-4bfa-831d-1e166fd89802 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.034742] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 904.034742] env[62521]: value = "task-1318720" [ 904.034742] env[62521]: _type = "Task" [ 904.034742] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.044462] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Created folder: Project (11829634203c4967bbcff90eb904097c) in parent group-v282025. [ 904.044756] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating folder: Instances. Parent ref: group-v282125. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 904.045094] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-390a17c3-ee98-4981-b132-67d0b4387063 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.050545] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318720, 'name': Rename_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.051621] env[62521]: DEBUG nova.network.neutron [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Updating instance_info_cache with network_info: [{"id": "2c0a3f29-63a4-44ff-bbf5-40908d17dc42", "address": "fa:16:3e:e1:0f:16", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c0a3f29-63", "ovs_interfaceid": "2c0a3f29-63a4-44ff-bbf5-40908d17dc42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.061274] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Created folder: Instances in parent group-v282125. [ 904.061604] env[62521]: DEBUG oslo.service.loopingcall [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.061868] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 904.062161] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3bb4c7e-08e7-4ed9-ab7e-4bd44b3ad6b3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.082351] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.082351] env[62521]: value = "task-1318723" [ 904.082351] env[62521]: _type = "Task" [ 904.082351] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.091120] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318723, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.279820] env[62521]: DEBUG nova.network.neutron [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Successfully created port: 4667150a-b284-4514-a637-498fbe7b40b3 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.310776] env[62521]: DEBUG oslo_vmware.api [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.326303] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 17876bcc-3a7b-47b0-afd9-1215e24864ae has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 904.326591] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 904.326719] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3712MB phys_disk=200GB used_disk=16GB total_vcpus=48 used_vcpus=16 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 904.451298] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "3805d897-fcb5-43ed-b502-7fad71124e74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.452482] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "3805d897-fcb5-43ed-b502-7fad71124e74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.452482] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "3805d897-fcb5-43ed-b502-7fad71124e74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.545064] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318720, 'name': Rename_Task, 'duration_secs': 0.230401} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.545417] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.545826] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9352bae9-9269-41f1-8f7f-b4757e268f38 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.552379] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 904.552379] env[62521]: value = "task-1318724" [ 904.552379] env[62521]: _type = "Task" [ 904.552379] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.555352] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "refresh_cache-563ad3b2-e53e-4682-9cde-bd6f709718de" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.555716] env[62521]: DEBUG nova.compute.manager [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Instance network_info: |[{"id": "2c0a3f29-63a4-44ff-bbf5-40908d17dc42", "address": "fa:16:3e:e1:0f:16", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c0a3f29-63", "ovs_interfaceid": "2c0a3f29-63a4-44ff-bbf5-40908d17dc42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 904.559316] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:0f:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd998416-f3d6-4a62-b828-5011063ce76a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c0a3f29-63a4-44ff-bbf5-40908d17dc42', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.566887] env[62521]: DEBUG oslo.service.loopingcall [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.571769] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 904.572617] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-311ca97e-62b0-4691-a66a-27563b1f156e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.594121] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318724, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.603734] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318723, 'name': CreateVM_Task, 'duration_secs': 0.492164} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.604918] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 904.605499] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.605499] env[62521]: value = "task-1318725" [ 904.605499] env[62521]: _type = "Task" [ 904.605499] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.606570] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.606716] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.607095] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.607801] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-495d7254-7b9e-4de6-9818-59a75442cd83 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.619316] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 904.619316] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52867016-db70-ebe0-1be6-899e32159e8b" [ 904.619316] env[62521]: _type = "Task" [ 904.619316] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.623414] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318725, 'name': CreateVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.635731] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52867016-db70-ebe0-1be6-899e32159e8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.660310] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ee2081-ae73-4b45-8436-a1c06b6e849d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.671011] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c94bf6-3fee-4038-9fcd-048251eb5cf4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.709022] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30dd9ac-98f4-44e5-9101-1d1a3c8262a8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.716514] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5ec468-ba0c-4dab-86b5-810894667a9c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.731530] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.771636] env[62521]: DEBUG nova.compute.manager [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 904.808314] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.810323] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.810323] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.810323] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.810323] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.810323] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.810323] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.810323] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.810323] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.810323] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.810323] env[62521]: DEBUG nova.virt.hardware [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.811429] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f14ec66-3117-4f25-ba89-efdf220cc5d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.818518] env[62521]: DEBUG oslo_vmware.api [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.822747] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8348c17b-e926-49fc-be2c-94ef7f6646cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.826906] env[62521]: DEBUG nova.network.neutron [req-4564dd60-3528-42ab-9059-9215a0f0bb0f req-721e5971-8754-49c4-a33b-d693f5789a0c service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updated VIF entry in instance network info cache for port 9e954d34-4424-459c-a022-fc039668c616. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 904.827267] env[62521]: DEBUG nova.network.neutron [req-4564dd60-3528-42ab-9059-9215a0f0bb0f req-721e5971-8754-49c4-a33b-d693f5789a0c service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance_info_cache with network_info: [{"id": "9e954d34-4424-459c-a022-fc039668c616", "address": "fa:16:3e:90:66:a0", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e954d34-44", "ovs_interfaceid": "9e954d34-4424-459c-a022-fc039668c616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.866071] env[62521]: DEBUG nova.network.neutron [-] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.933787] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "e2104c11-7713-4542-83da-bc4d2534113a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.934223] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "e2104c11-7713-4542-83da-bc4d2534113a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.934477] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "e2104c11-7713-4542-83da-bc4d2534113a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.935839] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "e2104c11-7713-4542-83da-bc4d2534113a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.935839] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "e2104c11-7713-4542-83da-bc4d2534113a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.937559] env[62521]: INFO nova.compute.manager [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Terminating instance [ 904.942562] env[62521]: DEBUG nova.compute.manager [req-d5eefa1d-e2b5-421e-a3b4-cecec734f634 req-c0b543c8-f7a9-4037-8fc2-63c5b27295a0 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Received event network-changed-2c0a3f29-63a4-44ff-bbf5-40908d17dc42 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.943118] env[62521]: DEBUG nova.compute.manager [req-d5eefa1d-e2b5-421e-a3b4-cecec734f634 req-c0b543c8-f7a9-4037-8fc2-63c5b27295a0 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Refreshing instance network info cache due to event network-changed-2c0a3f29-63a4-44ff-bbf5-40908d17dc42. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 904.944036] env[62521]: DEBUG oslo_concurrency.lockutils [req-d5eefa1d-e2b5-421e-a3b4-cecec734f634 req-c0b543c8-f7a9-4037-8fc2-63c5b27295a0 service nova] Acquiring lock "refresh_cache-563ad3b2-e53e-4682-9cde-bd6f709718de" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.944036] env[62521]: DEBUG oslo_concurrency.lockutils [req-d5eefa1d-e2b5-421e-a3b4-cecec734f634 req-c0b543c8-f7a9-4037-8fc2-63c5b27295a0 service nova] Acquired lock "refresh_cache-563ad3b2-e53e-4682-9cde-bd6f709718de" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.944311] env[62521]: DEBUG nova.network.neutron [req-d5eefa1d-e2b5-421e-a3b4-cecec734f634 req-c0b543c8-f7a9-4037-8fc2-63c5b27295a0 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Refreshing network info cache for port 2c0a3f29-63a4-44ff-bbf5-40908d17dc42 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.946751] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "refresh_cache-e2104c11-7713-4542-83da-bc4d2534113a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.946999] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquired lock "refresh_cache-e2104c11-7713-4542-83da-bc4d2534113a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.947268] env[62521]: DEBUG nova.network.neutron [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 905.066545] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318724, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.117727] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318725, 'name': CreateVM_Task, 'duration_secs': 0.37742} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.117904] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 905.118673] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.118853] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.119211] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 905.119572] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9ca6ba3-55f8-46df-b21f-ef02a47b99d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.124401] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 905.124401] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a8df8b-1daa-ddc6-c40c-6d597ad56813" [ 905.124401] env[62521]: _type = "Task" [ 905.124401] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.137584] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52867016-db70-ebe0-1be6-899e32159e8b, 'name': SearchDatastore_Task, 'duration_secs': 0.016638} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.140286] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.140506] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.140741] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.140890] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.141084] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.141398] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a8df8b-1daa-ddc6-c40c-6d597ad56813, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.141625] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef6d2c74-a41f-4676-a454-c2072c7aa4ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.149751] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.149947] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 905.150699] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ec747bc-5187-43c5-8f49-7b18d723cac4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.157615] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 905.157615] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5263d98f-8bc8-1f67-35c6-2d0aede31d95" [ 905.157615] env[62521]: _type = "Task" [ 905.157615] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.165031] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5263d98f-8bc8-1f67-35c6-2d0aede31d95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.234544] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.311693] env[62521]: DEBUG oslo_vmware.api [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318719, 'name': PowerOffVM_Task, 'duration_secs': 1.17583} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.311693] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 905.312120] env[62521]: DEBUG nova.compute.manager [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.312587] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94304bbc-948b-44f6-8a92-69f324f19918 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.330332] env[62521]: DEBUG oslo_concurrency.lockutils [req-4564dd60-3528-42ab-9059-9215a0f0bb0f req-721e5971-8754-49c4-a33b-d693f5789a0c service nova] Releasing lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.368831] env[62521]: INFO nova.compute.manager [-] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Took 1.36 seconds to deallocate network for instance. [ 905.501873] env[62521]: DEBUG nova.network.neutron [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 905.562739] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.562739] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.562739] env[62521]: DEBUG nova.network.neutron [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 905.569634] env[62521]: DEBUG oslo_vmware.api [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318724, 'name': PowerOnVM_Task, 'duration_secs': 0.534668} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.569760] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.570065] env[62521]: INFO nova.compute.manager [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Took 11.68 seconds to spawn the instance on the hypervisor. [ 905.570294] env[62521]: DEBUG nova.compute.manager [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.571258] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e774be56-2d48-4cf3-9e4c-549b64e3b94b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.622922] env[62521]: DEBUG nova.network.neutron [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.643066] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a8df8b-1daa-ddc6-c40c-6d597ad56813, 'name': SearchDatastore_Task, 'duration_secs': 0.011128} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.644061] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.644346] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.644601] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.644755] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.644940] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.645626] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ebd29b43-f410-417f-ba18-3cd97bca76d9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.655340] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.655548] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 905.656395] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-846bc758-dbca-4287-9ecd-8a51483e80de {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.665249] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 905.665249] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bb1d31-d1c7-aeea-65ec-6cbcc872e93e" [ 905.665249] env[62521]: _type = "Task" [ 905.665249] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.669344] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5263d98f-8bc8-1f67-35c6-2d0aede31d95, 'name': SearchDatastore_Task, 'duration_secs': 0.013969} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.673974] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5272b732-22a4-408a-9f2e-b6138d7a088e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.683334] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bb1d31-d1c7-aeea-65ec-6cbcc872e93e, 'name': SearchDatastore_Task, 'duration_secs': 0.014073} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.683772] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 905.683772] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5284dd4d-240e-bfb3-e6bd-16ef75228a78" [ 905.683772] env[62521]: _type = "Task" [ 905.683772] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.684661] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6df549f-6817-422c-ae01-7c4e9223e71b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.693947] env[62521]: DEBUG nova.network.neutron [req-d5eefa1d-e2b5-421e-a3b4-cecec734f634 req-c0b543c8-f7a9-4037-8fc2-63c5b27295a0 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Updated VIF entry in instance network info cache for port 2c0a3f29-63a4-44ff-bbf5-40908d17dc42. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 905.694358] env[62521]: DEBUG nova.network.neutron [req-d5eefa1d-e2b5-421e-a3b4-cecec734f634 req-c0b543c8-f7a9-4037-8fc2-63c5b27295a0 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Updating instance_info_cache with network_info: [{"id": "2c0a3f29-63a4-44ff-bbf5-40908d17dc42", "address": "fa:16:3e:e1:0f:16", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c0a3f29-63", "ovs_interfaceid": "2c0a3f29-63a4-44ff-bbf5-40908d17dc42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.695688] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 905.695688] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b7d061-5d56-0b94-12b0-8bc13c1be873" [ 905.695688] env[62521]: _type = "Task" [ 905.695688] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.702015] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5284dd4d-240e-bfb3-e6bd-16ef75228a78, 'name': SearchDatastore_Task, 'duration_secs': 0.011995} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.702858] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.703170] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 11e61e72-a311-4c43-bf53-df82ad9d70f7/11e61e72-a311-4c43-bf53-df82ad9d70f7.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 905.703509] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad125a2d-42b5-4efe-b430-ffddfd2b20e8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.712073] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b7d061-5d56-0b94-12b0-8bc13c1be873, 'name': SearchDatastore_Task, 'duration_secs': 0.012139} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.713401] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.713707] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 563ad3b2-e53e-4682-9cde-bd6f709718de/563ad3b2-e53e-4682-9cde-bd6f709718de.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 905.714145] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 905.714145] env[62521]: value = "task-1318726" [ 905.714145] env[62521]: _type = "Task" [ 905.714145] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.714466] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d0fdcca-a419-4edf-9528-bfe66c7d7741 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.726494] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318726, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.727763] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 905.727763] env[62521]: value = "task-1318727" [ 905.727763] env[62521]: _type = "Task" [ 905.727763] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.735789] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318727, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.739561] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 905.739762] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.978s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.740050] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.528s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.741594] env[62521]: INFO nova.compute.claims [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.825037] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e48460ce-e179-4d54-a796-f188ab76fe60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.559s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.875214] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.099411] env[62521]: INFO nova.compute.manager [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Took 32.01 seconds to build instance. [ 906.136389] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Releasing lock "refresh_cache-e2104c11-7713-4542-83da-bc4d2534113a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.136893] env[62521]: DEBUG nova.compute.manager [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 906.137109] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.139419] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b2b05b-6f4d-40d8-8b33-b096f4f0d570 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.147430] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.147912] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc6c8b74-14ba-4221-864a-5f1cbfe8e64a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.157902] env[62521]: DEBUG oslo_vmware.api [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 906.157902] env[62521]: value = "task-1318728" [ 906.157902] env[62521]: _type = "Task" [ 906.157902] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.169509] env[62521]: DEBUG oslo_vmware.api [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318728, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.201227] env[62521]: DEBUG oslo_concurrency.lockutils [req-d5eefa1d-e2b5-421e-a3b4-cecec734f634 req-c0b543c8-f7a9-4037-8fc2-63c5b27295a0 service nova] Releasing lock "refresh_cache-563ad3b2-e53e-4682-9cde-bd6f709718de" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.201624] env[62521]: DEBUG nova.compute.manager [req-d5eefa1d-e2b5-421e-a3b4-cecec734f634 req-c0b543c8-f7a9-4037-8fc2-63c5b27295a0 service nova] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Received event network-vif-deleted-ecf8506b-ea0e-48e2-8868-af841b45fe97 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.227120] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318726, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.242910] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318727, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.371704] env[62521]: DEBUG nova.compute.manager [req-ed2497d2-f565-4ae4-a714-49fc4b854977 req-6ddf4b18-7b8a-4e3d-826d-af0011906f6c service nova] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Received event network-vif-plugged-4667150a-b284-4514-a637-498fbe7b40b3 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.372076] env[62521]: DEBUG oslo_concurrency.lockutils [req-ed2497d2-f565-4ae4-a714-49fc4b854977 req-6ddf4b18-7b8a-4e3d-826d-af0011906f6c service nova] Acquiring lock "611462d2-7f57-44b0-ac36-db32af3d0dd0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.373178] env[62521]: DEBUG oslo_concurrency.lockutils [req-ed2497d2-f565-4ae4-a714-49fc4b854977 req-6ddf4b18-7b8a-4e3d-826d-af0011906f6c service nova] Lock "611462d2-7f57-44b0-ac36-db32af3d0dd0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.373178] env[62521]: DEBUG oslo_concurrency.lockutils [req-ed2497d2-f565-4ae4-a714-49fc4b854977 req-6ddf4b18-7b8a-4e3d-826d-af0011906f6c service nova] Lock "611462d2-7f57-44b0-ac36-db32af3d0dd0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.373178] env[62521]: DEBUG nova.compute.manager [req-ed2497d2-f565-4ae4-a714-49fc4b854977 req-6ddf4b18-7b8a-4e3d-826d-af0011906f6c service nova] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] No waiting events found dispatching network-vif-plugged-4667150a-b284-4514-a637-498fbe7b40b3 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.373528] env[62521]: WARNING nova.compute.manager [req-ed2497d2-f565-4ae4-a714-49fc4b854977 req-6ddf4b18-7b8a-4e3d-826d-af0011906f6c service nova] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Received unexpected event network-vif-plugged-4667150a-b284-4514-a637-498fbe7b40b3 for instance with vm_state building and task_state spawning. [ 906.375972] env[62521]: DEBUG nova.network.neutron [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance_info_cache with network_info: [{"id": "1643da59-fa10-4645-994b-0c557c3a6e10", "address": "fa:16:3e:11:0f:18", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1643da59-fa", "ovs_interfaceid": "1643da59-fa10-4645-994b-0c557c3a6e10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.418483] env[62521]: DEBUG oslo_concurrency.lockutils [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "6c4c36b7-b242-4578-9c23-d2529f308cb1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.601386] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6cd5e74e-f74e-474d-b973-a5b1e816788a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "6c4c36b7-b242-4578-9c23-d2529f308cb1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.525s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.601736] env[62521]: DEBUG oslo_concurrency.lockutils [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "6c4c36b7-b242-4578-9c23-d2529f308cb1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.183s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.601971] env[62521]: DEBUG oslo_concurrency.lockutils [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "6c4c36b7-b242-4578-9c23-d2529f308cb1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.602223] env[62521]: DEBUG oslo_concurrency.lockutils [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "6c4c36b7-b242-4578-9c23-d2529f308cb1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.602415] env[62521]: DEBUG oslo_concurrency.lockutils [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "6c4c36b7-b242-4578-9c23-d2529f308cb1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.605160] env[62521]: INFO nova.compute.manager [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Terminating instance [ 906.607209] env[62521]: DEBUG nova.compute.manager [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 906.607436] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.608336] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85a6176-ecb0-4492-8877-1de3e2c79ae5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.616679] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.617019] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54b325f2-10ee-4557-9e6a-78c8c8c16888 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.624071] env[62521]: DEBUG oslo_vmware.api [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 906.624071] env[62521]: value = "task-1318729" [ 906.624071] env[62521]: _type = "Task" [ 906.624071] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.634642] env[62521]: DEBUG oslo_vmware.api [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.667327] env[62521]: DEBUG oslo_vmware.api [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318728, 'name': PowerOffVM_Task, 'duration_secs': 0.197942} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.667524] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.667614] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.667947] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a77b9c5b-b4d2-4f8d-be2e-e2f44cb8f09f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.695693] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.696117] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.696535] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Deleting the datastore file [datastore2] e2104c11-7713-4542-83da-bc4d2534113a {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.696925] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a5938d8-444c-4129-a38e-ec89d992f051 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.704770] env[62521]: DEBUG oslo_vmware.api [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for the task: (returnval){ [ 906.704770] env[62521]: value = "task-1318731" [ 906.704770] env[62521]: _type = "Task" [ 906.704770] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.713286] env[62521]: DEBUG oslo_vmware.api [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.726930] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318726, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.733023} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.727221] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 11e61e72-a311-4c43-bf53-df82ad9d70f7/11e61e72-a311-4c43-bf53-df82ad9d70f7.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 906.727460] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.727729] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fa19667-8828-43c0-9365-0ca2fecf0ae0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.735179] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 906.735179] env[62521]: value = "task-1318732" [ 906.735179] env[62521]: _type = "Task" [ 906.735179] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.741800] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318727, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.801147} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.743090] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 563ad3b2-e53e-4682-9cde-bd6f709718de/563ad3b2-e53e-4682-9cde-bd6f709718de.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 906.743090] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.743090] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30491210-ce75-458d-bbe1-4db323c34d06 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.751479] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318732, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.756016] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 906.756016] env[62521]: value = "task-1318733" [ 906.756016] env[62521]: _type = "Task" [ 906.756016] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.764803] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318733, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.829891] env[62521]: DEBUG nova.network.neutron [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Successfully updated port: 4667150a-b284-4514-a637-498fbe7b40b3 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.881014] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.947976] env[62521]: DEBUG nova.compute.manager [req-5a817afe-7954-412f-bbda-1ebce30dbfb2 req-0f42166c-f722-4359-8129-03bdb64148fa service nova] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Received event network-changed-4667150a-b284-4514-a637-498fbe7b40b3 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.948245] env[62521]: DEBUG nova.compute.manager [req-5a817afe-7954-412f-bbda-1ebce30dbfb2 req-0f42166c-f722-4359-8129-03bdb64148fa service nova] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Refreshing instance network info cache due to event network-changed-4667150a-b284-4514-a637-498fbe7b40b3. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.948479] env[62521]: DEBUG oslo_concurrency.lockutils [req-5a817afe-7954-412f-bbda-1ebce30dbfb2 req-0f42166c-f722-4359-8129-03bdb64148fa service nova] Acquiring lock "refresh_cache-611462d2-7f57-44b0-ac36-db32af3d0dd0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.948757] env[62521]: DEBUG oslo_concurrency.lockutils [req-5a817afe-7954-412f-bbda-1ebce30dbfb2 req-0f42166c-f722-4359-8129-03bdb64148fa service nova] Acquired lock "refresh_cache-611462d2-7f57-44b0-ac36-db32af3d0dd0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.948951] env[62521]: DEBUG nova.network.neutron [req-5a817afe-7954-412f-bbda-1ebce30dbfb2 req-0f42166c-f722-4359-8129-03bdb64148fa service nova] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Refreshing network info cache for port 4667150a-b284-4514-a637-498fbe7b40b3 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 907.038935] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb97989-438d-44c9-b2f0-1705cdc542b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.046877] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd9f52b-2e89-4001-b783-b0906a6d8890 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.079135] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a9fc5b-57de-4817-bf9b-b87898bc8cb1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.087511] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768a075a-3f72-424a-8634-8b617536817e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.102171] env[62521]: DEBUG nova.compute.provider_tree [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.125751] env[62521]: DEBUG nova.objects.instance [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'flavor' on Instance uuid 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.137509] env[62521]: DEBUG oslo_vmware.api [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318729, 'name': PowerOffVM_Task, 'duration_secs': 0.191517} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.137837] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 907.138082] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 907.138382] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7410229-b17e-4797-943d-d007864e344a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.215106] env[62521]: DEBUG oslo_vmware.api [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Task: {'id': task-1318731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17098} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.215421] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.215651] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 907.215872] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 907.216033] env[62521]: INFO nova.compute.manager [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Took 1.08 seconds to destroy the instance on the hypervisor. [ 907.216333] env[62521]: DEBUG oslo.service.loopingcall [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.216557] env[62521]: DEBUG nova.compute.manager [-] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.216654] env[62521]: DEBUG nova.network.neutron [-] [instance: e2104c11-7713-4542-83da-bc4d2534113a] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 907.231207] env[62521]: DEBUG nova.network.neutron [-] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.246523] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318732, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113831} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.246826] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.247709] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51059f52-3f32-42fd-b430-2fa38835a64d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.253595] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 907.253909] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 907.254216] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Deleting the datastore file [datastore2] 6c4c36b7-b242-4578-9c23-d2529f308cb1 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 907.263374] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bda1171f-5d2c-48aa-9e93-952dd12890ec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.275108] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 11e61e72-a311-4c43-bf53-df82ad9d70f7/11e61e72-a311-4c43-bf53-df82ad9d70f7.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.277598] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63fd8393-7454-4079-aad2-b3e67f98b080 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.297440] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318733, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061915} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.298739] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.299108] env[62521]: DEBUG oslo_vmware.api [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for the task: (returnval){ [ 907.299108] env[62521]: value = "task-1318735" [ 907.299108] env[62521]: _type = "Task" [ 907.299108] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.300676] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d89a9e-14b9-4122-a6fd-14511e885b74 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.303620] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 907.303620] env[62521]: value = "task-1318736" [ 907.303620] env[62521]: _type = "Task" [ 907.303620] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.331399] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 563ad3b2-e53e-4682-9cde-bd6f709718de/563ad3b2-e53e-4682-9cde-bd6f709718de.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.335900] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "refresh_cache-611462d2-7f57-44b0-ac36-db32af3d0dd0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.336183] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52b31aef-bd6a-4624-8b93-76a60b28cd14 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.351486] env[62521]: DEBUG oslo_vmware.api [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318735, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.355136] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318736, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.360482] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 907.360482] env[62521]: value = "task-1318737" [ 907.360482] env[62521]: _type = "Task" [ 907.360482] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.370493] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318737, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.403293] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe18c7a-7b35-4742-a55f-9ca22c7e13a4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.423598] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c02ffb-2e39-4bfe-a99b-4a13f4574f2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.431636] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance '3805d897-fcb5-43ed-b502-7fad71124e74' progress to 83 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 907.511306] env[62521]: DEBUG nova.network.neutron [req-5a817afe-7954-412f-bbda-1ebce30dbfb2 req-0f42166c-f722-4359-8129-03bdb64148fa service nova] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.613031] env[62521]: DEBUG nova.scheduler.client.report [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.631785] env[62521]: DEBUG oslo_concurrency.lockutils [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.631995] env[62521]: DEBUG oslo_concurrency.lockutils [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.632201] env[62521]: DEBUG nova.network.neutron [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.632440] env[62521]: DEBUG nova.objects.instance [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'info_cache' on Instance uuid 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.682112] env[62521]: DEBUG nova.network.neutron [req-5a817afe-7954-412f-bbda-1ebce30dbfb2 req-0f42166c-f722-4359-8129-03bdb64148fa service nova] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.734763] env[62521]: DEBUG nova.network.neutron [-] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.821826] env[62521]: DEBUG oslo_vmware.api [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Task: {'id': task-1318735, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198951} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.824304] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.824515] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 907.824701] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 907.824878] env[62521]: INFO nova.compute.manager [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Took 1.22 seconds to destroy the instance on the hypervisor. [ 907.825139] env[62521]: DEBUG oslo.service.loopingcall [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.825426] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318736, 'name': ReconfigVM_Task, 'duration_secs': 0.49275} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.825574] env[62521]: DEBUG nova.compute.manager [-] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.825667] env[62521]: DEBUG nova.network.neutron [-] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 907.827424] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 11e61e72-a311-4c43-bf53-df82ad9d70f7/11e61e72-a311-4c43-bf53-df82ad9d70f7.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.828063] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-871646e2-4670-412f-a862-e3a8f4005c36 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.834532] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 907.834532] env[62521]: value = "task-1318738" [ 907.834532] env[62521]: _type = "Task" [ 907.834532] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.842822] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318738, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.870178] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318737, 'name': ReconfigVM_Task, 'duration_secs': 0.411422} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.870444] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 563ad3b2-e53e-4682-9cde-bd6f709718de/563ad3b2-e53e-4682-9cde-bd6f709718de.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.871154] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57ff2516-209f-4935-a0d2-1971ba53d9e9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.877909] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 907.877909] env[62521]: value = "task-1318739" [ 907.877909] env[62521]: _type = "Task" [ 907.877909] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.886630] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318739, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.938946] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.939345] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3997db61-cd91-4773-b3e8-6b6227c71f88 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.947051] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 907.947051] env[62521]: value = "task-1318740" [ 907.947051] env[62521]: _type = "Task" [ 907.947051] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.957553] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318740, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.123019] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.123544] env[62521]: DEBUG nova.compute.manager [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.127773] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.371s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.128021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.131206] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 4.136s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.131407] env[62521]: DEBUG nova.objects.instance [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62521) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 908.139606] env[62521]: DEBUG nova.objects.base [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Object Instance<7b5afeab-f04e-453a-a3ed-09e3a725e8d6> lazy-loaded attributes: flavor,info_cache {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 908.167987] env[62521]: INFO nova.scheduler.client.report [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleted allocations for instance 1d1d2235-748a-4f27-a915-b69f9110257b [ 908.185883] env[62521]: DEBUG oslo_concurrency.lockutils [req-5a817afe-7954-412f-bbda-1ebce30dbfb2 req-0f42166c-f722-4359-8129-03bdb64148fa service nova] Releasing lock "refresh_cache-611462d2-7f57-44b0-ac36-db32af3d0dd0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.185883] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "refresh_cache-611462d2-7f57-44b0-ac36-db32af3d0dd0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.185883] env[62521]: DEBUG nova.network.neutron [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 908.235912] env[62521]: INFO nova.compute.manager [-] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Took 1.02 seconds to deallocate network for instance. [ 908.344966] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318738, 'name': Rename_Task, 'duration_secs': 0.178026} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.345823] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 908.346149] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce01dbe5-c848-4b67-a3b8-35c904450b44 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.352776] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 908.352776] env[62521]: value = "task-1318741" [ 908.352776] env[62521]: _type = "Task" [ 908.352776] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.360790] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.386878] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318739, 'name': Rename_Task, 'duration_secs': 0.166162} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.387748] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 908.388048] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b18c7f1-bc20-4740-af20-f78de441af83 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.394370] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 908.394370] env[62521]: value = "task-1318742" [ 908.394370] env[62521]: _type = "Task" [ 908.394370] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.402913] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318742, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.457424] env[62521]: DEBUG oslo_vmware.api [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318740, 'name': PowerOnVM_Task, 'duration_secs': 0.465664} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.457886] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.457886] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d99c4bea-0b26-4567-9903-c0fdcd53042d tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance '3805d897-fcb5-43ed-b502-7fad71124e74' progress to 100 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 908.637709] env[62521]: DEBUG nova.compute.utils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.639745] env[62521]: DEBUG nova.compute.manager [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.639969] env[62521]: DEBUG nova.network.neutron [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 908.679621] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04eddeb9-3267-4f36-b959-4cd11956ae62 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "1d1d2235-748a-4f27-a915-b69f9110257b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.097s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.717871] env[62521]: DEBUG nova.policy [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '937840982b914e68849319714a16b1a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea161b1af2bb4b9f8cc39b16582f7013', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 908.749340] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.752639] env[62521]: DEBUG nova.network.neutron [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 908.862815] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318741, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.905477] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318742, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.021904] env[62521]: DEBUG nova.network.neutron [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Successfully created port: 3b0934da-f42d-4e91-ae7c-a4f254d6b86c {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.143124] env[62521]: DEBUG nova.compute.manager [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.150518] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47a5b3aa-b3db-45b3-8a59-a788ad49e856 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.151676] env[62521]: DEBUG nova.network.neutron [-] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.156150] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.281s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.156150] env[62521]: DEBUG nova.objects.instance [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lazy-loading 'resources' on Instance uuid 3447712d-daa6-4329-8015-6474cca1a5e4 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.224451] env[62521]: DEBUG nova.compute.manager [req-5fb8af12-f6cd-4259-98b4-09b19ff950c1 req-90cc84c1-0653-49fc-847d-6e8167742cc9 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Received event network-vif-deleted-be96c060-332a-43b8-be9e-e62a6dcc3a9f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.224451] env[62521]: DEBUG nova.compute.manager [req-5fb8af12-f6cd-4259-98b4-09b19ff950c1 req-90cc84c1-0653-49fc-847d-6e8167742cc9 service nova] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Received event network-vif-deleted-c6583bbe-d9ae-4317-8a19-9163ed23ce1d {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.365218] env[62521]: DEBUG oslo_vmware.api [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318741, 'name': PowerOnVM_Task, 'duration_secs': 0.873904} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.365520] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.365839] env[62521]: INFO nova.compute.manager [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Took 9.38 seconds to spawn the instance on the hypervisor. [ 909.365925] env[62521]: DEBUG nova.compute.manager [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.366742] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca6d615-c73d-4750-87db-826631fed68e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.387934] env[62521]: DEBUG nova.network.neutron [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Updating instance_info_cache with network_info: [{"id": "4667150a-b284-4514-a637-498fbe7b40b3", "address": "fa:16:3e:cf:b4:a5", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4667150a-b2", "ovs_interfaceid": "4667150a-b284-4514-a637-498fbe7b40b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.404823] env[62521]: DEBUG oslo_vmware.api [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318742, 'name': PowerOnVM_Task, 'duration_secs': 0.834863} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.406606] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.406606] env[62521]: INFO nova.compute.manager [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Took 7.47 seconds to spawn the instance on the hypervisor. [ 909.406606] env[62521]: DEBUG nova.compute.manager [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.407414] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674b20f2-f8a9-4599-94f6-88c1ab3c485f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.419519] env[62521]: DEBUG nova.network.neutron [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [{"id": "8824a231-01df-41e5-9bea-9071d8b41cb8", "address": "fa:16:3e:2e:cc:1b", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8824a231-01", "ovs_interfaceid": "8824a231-01df-41e5-9bea-9071d8b41cb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.661980] env[62521]: INFO nova.compute.manager [-] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Took 1.84 seconds to deallocate network for instance. [ 909.686257] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "4dcd1c6c-8726-42db-997a-e78ba1293310" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.686584] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "4dcd1c6c-8726-42db-997a-e78ba1293310" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.686871] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "4dcd1c6c-8726-42db-997a-e78ba1293310-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.687176] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "4dcd1c6c-8726-42db-997a-e78ba1293310-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.687424] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "4dcd1c6c-8726-42db-997a-e78ba1293310-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.692972] env[62521]: INFO nova.compute.manager [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Terminating instance [ 909.698348] env[62521]: DEBUG nova.compute.manager [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.698617] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 909.700739] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24643826-43a3-4568-ba7e-874b5d7353f1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.710107] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.710362] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-19026500-8ac6-4699-a0ed-f8642f63be84 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.719324] env[62521]: DEBUG oslo_vmware.api [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 909.719324] env[62521]: value = "task-1318743" [ 909.719324] env[62521]: _type = "Task" [ 909.719324] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.727522] env[62521]: DEBUG oslo_vmware.api [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.886659] env[62521]: INFO nova.compute.manager [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Took 26.78 seconds to build instance. [ 909.890571] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "refresh_cache-611462d2-7f57-44b0-ac36-db32af3d0dd0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.890921] env[62521]: DEBUG nova.compute.manager [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Instance network_info: |[{"id": "4667150a-b284-4514-a637-498fbe7b40b3", "address": "fa:16:3e:cf:b4:a5", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4667150a-b2", "ovs_interfaceid": "4667150a-b284-4514-a637-498fbe7b40b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 909.891405] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:b4:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '84aee122-f630-43c5-9cc1-3a38d3819c82', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4667150a-b284-4514-a637-498fbe7b40b3', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 909.900089] env[62521]: DEBUG oslo.service.loopingcall [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.900657] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 909.900657] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b86e40d-fc61-4a69-ac05-bdfa6670deb5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.925685] env[62521]: DEBUG oslo_concurrency.lockutils [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.927421] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 909.927421] env[62521]: value = "task-1318744" [ 909.927421] env[62521]: _type = "Task" [ 909.927421] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.934604] env[62521]: INFO nova.compute.manager [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Took 24.69 seconds to build instance. [ 909.943537] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318744, 'name': CreateVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.963589] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b9aec3-5eac-49c2-b588-3eccb3971518 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.972528] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1edd79b-07a9-48d5-9461-5303c4a2ba70 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.015787] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3002da4-837e-47bf-abf5-c88119cd3dcc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.025150] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3f357e-9a7a-4616-8aaa-50b0365fc359 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.041153] env[62521]: DEBUG nova.compute.provider_tree [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.163694] env[62521]: DEBUG nova.compute.manager [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.168734] env[62521]: DEBUG oslo_concurrency.lockutils [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.206784] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:36:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='e986ebeb-2013-4e50-9994-d8d533ed7ed2',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-778228551',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.207120] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.207362] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.207628] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.207812] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.207960] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.208198] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.208367] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.208607] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.208837] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.209080] env[62521]: DEBUG nova.virt.hardware [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.210050] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f457c6a1-0e2b-497d-97eb-b1d90a9d1169 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.219386] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e0ace4-3e9d-49a1-9f99-edb2c1fb2fa7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.242025] env[62521]: DEBUG oslo_vmware.api [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318743, 'name': PowerOffVM_Task, 'duration_secs': 0.19292} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.242345] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.242589] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.242873] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-866ac643-d4a7-463f-b157-ae6e04952419 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.329852] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.330010] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.330210] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleting the datastore file [datastore1] 4dcd1c6c-8726-42db-997a-e78ba1293310 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.330620] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e686a0bd-0edb-46c3-814c-5bc6d6b857c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.337166] env[62521]: DEBUG oslo_vmware.api [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 910.337166] env[62521]: value = "task-1318746" [ 910.337166] env[62521]: _type = "Task" [ 910.337166] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.346797] env[62521]: DEBUG oslo_vmware.api [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318746, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.388655] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1c5e4a19-37a4-4183-8bc3-0abf251f812b tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.293s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.429261] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.429657] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d1e19b8-afe3-436b-997a-025d0d0ca08f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.437178] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5046c901-6f97-4abe-b29f-564a958793bf tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "563ad3b2-e53e-4682-9cde-bd6f709718de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.203s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.442682] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318744, 'name': CreateVM_Task, 'duration_secs': 0.381613} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.443884] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 910.444295] env[62521]: DEBUG oslo_vmware.api [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 910.444295] env[62521]: value = "task-1318747" [ 910.444295] env[62521]: _type = "Task" [ 910.444295] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.444948] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.445258] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.445523] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.445732] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-064ce8fc-72cd-49dd-9062-318773cc206a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.455158] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 910.455158] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527f8687-43d5-d49b-7dc5-421fa6b2292c" [ 910.455158] env[62521]: _type = "Task" [ 910.455158] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.458340] env[62521]: DEBUG oslo_vmware.api [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318747, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.468544] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527f8687-43d5-d49b-7dc5-421fa6b2292c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.546016] env[62521]: DEBUG nova.scheduler.client.report [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.670945] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "3afba9cd-f66e-44c5-b660-b4314f370901" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.671996] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "3afba9cd-f66e-44c5-b660-b4314f370901" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.849728] env[62521]: DEBUG oslo_vmware.api [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318746, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.270698} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.850182] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.850503] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.850835] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.853095] env[62521]: INFO nova.compute.manager [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Took 1.15 seconds to destroy the instance on the hypervisor. [ 910.853095] env[62521]: DEBUG oslo.service.loopingcall [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.853095] env[62521]: DEBUG nova.compute.manager [-] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.853095] env[62521]: DEBUG nova.network.neutron [-] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 910.947167] env[62521]: DEBUG nova.compute.manager [req-992f7557-49b1-48fa-bb99-64ea7f19957e req-43f6ae9e-1851-49db-961f-eb3826784b1a service nova] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Received event network-vif-plugged-3b0934da-f42d-4e91-ae7c-a4f254d6b86c {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.947167] env[62521]: DEBUG oslo_concurrency.lockutils [req-992f7557-49b1-48fa-bb99-64ea7f19957e req-43f6ae9e-1851-49db-961f-eb3826784b1a service nova] Acquiring lock "17876bcc-3a7b-47b0-afd9-1215e24864ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.947710] env[62521]: DEBUG oslo_concurrency.lockutils [req-992f7557-49b1-48fa-bb99-64ea7f19957e req-43f6ae9e-1851-49db-961f-eb3826784b1a service nova] Lock "17876bcc-3a7b-47b0-afd9-1215e24864ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.948115] env[62521]: DEBUG oslo_concurrency.lockutils [req-992f7557-49b1-48fa-bb99-64ea7f19957e req-43f6ae9e-1851-49db-961f-eb3826784b1a service nova] Lock "17876bcc-3a7b-47b0-afd9-1215e24864ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.949437] env[62521]: DEBUG nova.compute.manager [req-992f7557-49b1-48fa-bb99-64ea7f19957e req-43f6ae9e-1851-49db-961f-eb3826784b1a service nova] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] No waiting events found dispatching network-vif-plugged-3b0934da-f42d-4e91-ae7c-a4f254d6b86c {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 910.949780] env[62521]: WARNING nova.compute.manager [req-992f7557-49b1-48fa-bb99-64ea7f19957e req-43f6ae9e-1851-49db-961f-eb3826784b1a service nova] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Received unexpected event network-vif-plugged-3b0934da-f42d-4e91-ae7c-a4f254d6b86c for instance with vm_state building and task_state spawning. [ 910.966787] env[62521]: DEBUG oslo_vmware.api [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318747, 'name': PowerOnVM_Task, 'duration_secs': 0.473395} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.967828] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 910.968268] env[62521]: DEBUG nova.compute.manager [None req-addf1151-65fd-4423-ad22-b09184e600e3 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.969729] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0d91d6-c521-415e-8f71-d45a0009ea1c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.979471] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527f8687-43d5-d49b-7dc5-421fa6b2292c, 'name': SearchDatastore_Task, 'duration_secs': 0.016131} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.979968] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.980228] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.980527] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.980731] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.980850] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.981785] env[62521]: DEBUG nova.network.neutron [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Successfully updated port: 3b0934da-f42d-4e91-ae7c-a4f254d6b86c {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.983216] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-72661c32-d9be-4e1e-ae3c-74a32e43751e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.004230] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.004517] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.006245] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d61e041-89b0-4da5-96ba-8e4f4d698534 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.012951] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 911.012951] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52425d47-9178-3246-6031-ee1abeca528b" [ 911.012951] env[62521]: _type = "Task" [ 911.012951] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.025645] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52425d47-9178-3246-6031-ee1abeca528b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.055295] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.899s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.059153] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.309s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.059425] env[62521]: DEBUG nova.objects.instance [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lazy-loading 'resources' on Instance uuid e2104c11-7713-4542-83da-bc4d2534113a {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.082944] env[62521]: INFO nova.scheduler.client.report [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Deleted allocations for instance 3447712d-daa6-4329-8015-6474cca1a5e4 [ 911.175296] env[62521]: DEBUG nova.compute.manager [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 911.334523] env[62521]: DEBUG nova.compute.manager [req-6f89c83b-1084-455f-aee7-1cea89172421 req-23d56bca-b757-43fe-9c89-9c9ac767f8c5 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Received event network-changed-2c0a3f29-63a4-44ff-bbf5-40908d17dc42 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.334733] env[62521]: DEBUG nova.compute.manager [req-6f89c83b-1084-455f-aee7-1cea89172421 req-23d56bca-b757-43fe-9c89-9c9ac767f8c5 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Refreshing instance network info cache due to event network-changed-2c0a3f29-63a4-44ff-bbf5-40908d17dc42. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.335473] env[62521]: DEBUG oslo_concurrency.lockutils [req-6f89c83b-1084-455f-aee7-1cea89172421 req-23d56bca-b757-43fe-9c89-9c9ac767f8c5 service nova] Acquiring lock "refresh_cache-563ad3b2-e53e-4682-9cde-bd6f709718de" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.335679] env[62521]: DEBUG oslo_concurrency.lockutils [req-6f89c83b-1084-455f-aee7-1cea89172421 req-23d56bca-b757-43fe-9c89-9c9ac767f8c5 service nova] Acquired lock "refresh_cache-563ad3b2-e53e-4682-9cde-bd6f709718de" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.335948] env[62521]: DEBUG nova.network.neutron [req-6f89c83b-1084-455f-aee7-1cea89172421 req-23d56bca-b757-43fe-9c89-9c9ac767f8c5 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Refreshing network info cache for port 2c0a3f29-63a4-44ff-bbf5-40908d17dc42 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 911.407713] env[62521]: DEBUG oslo_vmware.rw_handles [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e4633-3ce5-0d2d-7a92-4d41e45e1ae1/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 911.409295] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2227002d-6de2-477b-93a9-7827e028df8b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.418394] env[62521]: DEBUG oslo_vmware.rw_handles [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e4633-3ce5-0d2d-7a92-4d41e45e1ae1/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 911.418677] env[62521]: ERROR oslo_vmware.rw_handles [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e4633-3ce5-0d2d-7a92-4d41e45e1ae1/disk-0.vmdk due to incomplete transfer. [ 911.419035] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6bb0b830-b502-4206-8084-de9acb2cbd4c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.429517] env[62521]: DEBUG oslo_vmware.rw_handles [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e4633-3ce5-0d2d-7a92-4d41e45e1ae1/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 911.429731] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Uploaded image 720eb7fe-5500-4cf9-a06e-3471ca9179ed to the Glance image server {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 911.432015] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 911.432516] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "3805d897-fcb5-43ed-b502-7fad71124e74" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.432777] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "3805d897-fcb5-43ed-b502-7fad71124e74" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.432983] env[62521]: DEBUG nova.compute.manager [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Going to confirm migration 2 {{(pid=62521) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 911.434390] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-55f5848b-119c-4a80-aa61-27a88864e2f2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.442852] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 911.442852] env[62521]: value = "task-1318748" [ 911.442852] env[62521]: _type = "Task" [ 911.442852] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.451565] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318748, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.488097] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.488223] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.489917] env[62521]: DEBUG nova.network.neutron [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.529023] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52425d47-9178-3246-6031-ee1abeca528b, 'name': SearchDatastore_Task, 'duration_secs': 0.013396} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.529023] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ce89676-2e32-47ae-a701-58f772b94118 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.533383] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 911.533383] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f68b32-6ecb-2715-40f5-68fa987981a3" [ 911.533383] env[62521]: _type = "Task" [ 911.533383] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.541279] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f68b32-6ecb-2715-40f5-68fa987981a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.591164] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab9721fd-a6ed-4bdd-854c-36974ce42eeb tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "3447712d-daa6-4329-8015-6474cca1a5e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.844s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.694734] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.803039] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9556e6b4-e786-4d99-9139-8b3d829a8010 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.811209] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545bc65a-774b-4467-b7a1-f59164ffed40 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.844956] env[62521]: DEBUG nova.network.neutron [-] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.847534] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d487d2-28b1-4d90-a885-1f0d74a2f773 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.855487] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b09a7b6-e098-4197-8174-96b5434116bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.870365] env[62521]: DEBUG nova.compute.provider_tree [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.953575] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318748, 'name': Destroy_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.034386] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.034689] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.034906] env[62521]: DEBUG nova.network.neutron [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.035184] env[62521]: DEBUG nova.objects.instance [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lazy-loading 'info_cache' on Instance uuid 3805d897-fcb5-43ed-b502-7fad71124e74 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.054807] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f68b32-6ecb-2715-40f5-68fa987981a3, 'name': SearchDatastore_Task, 'duration_secs': 0.05343} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.054807] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.054807] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 611462d2-7f57-44b0-ac36-db32af3d0dd0/611462d2-7f57-44b0-ac36-db32af3d0dd0.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.054807] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7139c9ff-f529-47dc-8590-9246865a5fb1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.066026] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 912.066026] env[62521]: value = "task-1318749" [ 912.066026] env[62521]: _type = "Task" [ 912.066026] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.077525] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318749, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.096542] env[62521]: DEBUG nova.network.neutron [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 912.348104] env[62521]: INFO nova.compute.manager [-] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Took 1.50 seconds to deallocate network for instance. [ 912.375090] env[62521]: DEBUG nova.scheduler.client.report [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.458519] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318748, 'name': Destroy_Task, 'duration_secs': 0.660861} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.458958] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Destroyed the VM [ 912.459396] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 912.459795] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8f02c92c-b1e8-4403-afce-19ec1c5cd4f6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.467750] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 912.467750] env[62521]: value = "task-1318750" [ 912.467750] env[62521]: _type = "Task" [ 912.467750] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.479967] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318750, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.576656] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318749, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.680753] env[62521]: DEBUG nova.network.neutron [req-6f89c83b-1084-455f-aee7-1cea89172421 req-23d56bca-b757-43fe-9c89-9c9ac767f8c5 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Updated VIF entry in instance network info cache for port 2c0a3f29-63a4-44ff-bbf5-40908d17dc42. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 912.681196] env[62521]: DEBUG nova.network.neutron [req-6f89c83b-1084-455f-aee7-1cea89172421 req-23d56bca-b757-43fe-9c89-9c9ac767f8c5 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Updating instance_info_cache with network_info: [{"id": "2c0a3f29-63a4-44ff-bbf5-40908d17dc42", "address": "fa:16:3e:e1:0f:16", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c0a3f29-63", "ovs_interfaceid": "2c0a3f29-63a4-44ff-bbf5-40908d17dc42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.704257] env[62521]: DEBUG nova.network.neutron [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance_info_cache with network_info: [{"id": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "address": "fa:16:3e:1a:80:af", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0934da-f4", "ovs_interfaceid": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.855175] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.883783] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.826s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.887352] env[62521]: DEBUG oslo_concurrency.lockutils [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.719s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.887625] env[62521]: DEBUG nova.objects.instance [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lazy-loading 'resources' on Instance uuid 6c4c36b7-b242-4578-9c23-d2529f308cb1 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.917299] env[62521]: INFO nova.scheduler.client.report [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Deleted allocations for instance e2104c11-7713-4542-83da-bc4d2534113a [ 912.978867] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318750, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.074351] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318749, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.810629} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.074641] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 611462d2-7f57-44b0-ac36-db32af3d0dd0/611462d2-7f57-44b0-ac36-db32af3d0dd0.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 913.074860] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 913.075143] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff22dd37-c0ec-462c-b57b-168ab5a3dae2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.081676] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 913.081676] env[62521]: value = "task-1318751" [ 913.081676] env[62521]: _type = "Task" [ 913.081676] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.090213] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318751, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.184631] env[62521]: DEBUG oslo_concurrency.lockutils [req-6f89c83b-1084-455f-aee7-1cea89172421 req-23d56bca-b757-43fe-9c89-9c9ac767f8c5 service nova] Releasing lock "refresh_cache-563ad3b2-e53e-4682-9cde-bd6f709718de" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.207107] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.207485] env[62521]: DEBUG nova.compute.manager [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Instance network_info: |[{"id": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "address": "fa:16:3e:1a:80:af", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0934da-f4", "ovs_interfaceid": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 913.207940] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:80:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc7aa55d-223a-4157-9137-88dc492f2db2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3b0934da-f42d-4e91-ae7c-a4f254d6b86c', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.215851] env[62521]: DEBUG oslo.service.loopingcall [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.218451] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 913.218825] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c3cd1b0-cdbd-4fb8-a868-12cdd4b809f2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.240415] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.240415] env[62521]: value = "task-1318752" [ 913.240415] env[62521]: _type = "Task" [ 913.240415] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.248608] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318752, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.290886] env[62521]: DEBUG nova.compute.manager [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Received event network-changed-3b0934da-f42d-4e91-ae7c-a4f254d6b86c {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.291207] env[62521]: DEBUG nova.compute.manager [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Refreshing instance network info cache due to event network-changed-3b0934da-f42d-4e91-ae7c-a4f254d6b86c. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 913.291524] env[62521]: DEBUG oslo_concurrency.lockutils [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] Acquiring lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.291720] env[62521]: DEBUG oslo_concurrency.lockutils [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] Acquired lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.291943] env[62521]: DEBUG nova.network.neutron [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Refreshing network info cache for port 3b0934da-f42d-4e91-ae7c-a4f254d6b86c {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 913.396234] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.396488] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.425934] env[62521]: DEBUG oslo_concurrency.lockutils [None req-32d49b06-a733-4353-9dd6-f5a2e67b42a5 tempest-ServerShowV254Test-1657145180 tempest-ServerShowV254Test-1657145180-project-member] Lock "e2104c11-7713-4542-83da-bc4d2534113a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.492s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.484617] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318750, 'name': RemoveSnapshot_Task, 'duration_secs': 0.657493} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.489381] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 913.490115] env[62521]: DEBUG nova.compute.manager [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.491432] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b431fa89-217c-4432-a3f8-47d48cde1fe4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.507868] env[62521]: DEBUG nova.network.neutron [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance_info_cache with network_info: [{"id": "1643da59-fa10-4645-994b-0c557c3a6e10", "address": "fa:16:3e:11:0f:18", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1643da59-fa", "ovs_interfaceid": "1643da59-fa10-4645-994b-0c557c3a6e10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.543562] env[62521]: DEBUG nova.compute.manager [req-50bf433b-b02e-4632-8b18-2d4d32852ee2 req-4d2b7e47-d91c-404f-908d-73d1c248655f service nova] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Received event network-vif-deleted-12ed6066-6ad4-42ab-93ff-b25c2624d917 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.577531] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.577791] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.593443] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318751, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08121} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.597041] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.598941] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c4cd0a-900d-4970-a6da-97b7cd3e040f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.625179] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 611462d2-7f57-44b0-ac36-db32af3d0dd0/611462d2-7f57-44b0-ac36-db32af3d0dd0.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.628577] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-310555a7-5cb8-4402-8a1c-0d9d8b947bc6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.651949] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 913.651949] env[62521]: value = "task-1318753" [ 913.651949] env[62521]: _type = "Task" [ 913.651949] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.664081] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318753, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.710623] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9df5fd-8445-4f3b-8ffa-45c016890738 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.723617] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2b04bd-2472-4255-86a6-96d08c99e7df {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.760132] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746f4c26-b5b6-4f3e-aa36-ee0ad9e6e6cf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.771048] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318752, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.773697] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3b0eed-ed90-4a0a-8ee9-5bbd57646973 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.789350] env[62521]: DEBUG nova.compute.provider_tree [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.899773] env[62521]: INFO nova.compute.manager [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Detaching volume 6e8eee93-7abc-4c71-8667-16b6b4dbd033 [ 913.948461] env[62521]: INFO nova.virt.block_device [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Attempting to driver detach volume 6e8eee93-7abc-4c71-8667-16b6b4dbd033 from mountpoint /dev/sdb [ 913.948732] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 913.948930] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282105', 'volume_id': '6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'name': 'volume-6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '538ed18e-920e-4ed4-a9e1-a6bfbae7f282', 'attached_at': '', 'detached_at': '', 'volume_id': '6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'serial': '6e8eee93-7abc-4c71-8667-16b6b4dbd033'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 913.949868] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c6d041-033f-409a-87b2-43be90e3ff53 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.978319] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dbfb56d-59d0-43f8-9c52-79c107d16560 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.985764] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251c9c18-a0f2-4290-8dbd-e50779fa9e05 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.014869] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-3805d897-fcb5-43ed-b502-7fad71124e74" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.015168] env[62521]: DEBUG nova.objects.instance [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lazy-loading 'migration_context' on Instance uuid 3805d897-fcb5-43ed-b502-7fad71124e74 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.018309] env[62521]: INFO nova.compute.manager [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Shelve offloading [ 914.021024] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcad4bdb-1ea4-4224-9df7-1d09c4e7b206 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.024283] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.025195] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87dfe298-e0de-4851-9b0e-544245eb5595 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.042278] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] The volume has not been displaced from its original location: [datastore2] volume-6e8eee93-7abc-4c71-8667-16b6b4dbd033/volume-6e8eee93-7abc-4c71-8667-16b6b4dbd033.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 914.048357] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Reconfiguring VM instance instance-00000037 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 914.054108] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7b1bc32-c8a1-44c4-bdf1-11c2ba45a2ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.068733] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 914.068733] env[62521]: value = "task-1318754" [ 914.068733] env[62521]: _type = "Task" [ 914.068733] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.076103] env[62521]: DEBUG oslo_vmware.api [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 914.076103] env[62521]: value = "task-1318755" [ 914.076103] env[62521]: _type = "Task" [ 914.076103] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.082729] env[62521]: DEBUG nova.compute.manager [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 914.085845] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 914.086372] env[62521]: DEBUG nova.compute.manager [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.086995] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d850ecd-c4a1-4b79-88b8-0b058ad19e6d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.092781] env[62521]: DEBUG oslo_vmware.api [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318755, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.097850] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.097850] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.097850] env[62521]: DEBUG nova.network.neutron [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.162093] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318753, 'name': ReconfigVM_Task, 'duration_secs': 0.371572} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.162413] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 611462d2-7f57-44b0-ac36-db32af3d0dd0/611462d2-7f57-44b0-ac36-db32af3d0dd0.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.163077] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46387bcd-84b1-4122-aa7d-0ac2472b0a2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.172418] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 914.172418] env[62521]: value = "task-1318756" [ 914.172418] env[62521]: _type = "Task" [ 914.172418] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.178991] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318756, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.212522] env[62521]: DEBUG nova.network.neutron [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updated VIF entry in instance network info cache for port 3b0934da-f42d-4e91-ae7c-a4f254d6b86c. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 914.213014] env[62521]: DEBUG nova.network.neutron [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance_info_cache with network_info: [{"id": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "address": "fa:16:3e:1a:80:af", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0934da-f4", "ovs_interfaceid": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.265804] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318752, 'name': CreateVM_Task, 'duration_secs': 0.61296} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.265989] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 914.266794] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.266974] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.267342] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.267631] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-777b1875-45ae-4a8a-8de9-2ca88db0b74d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.272390] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 914.272390] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f648ea-99d6-dd17-d198-2f32d17a04a8" [ 914.272390] env[62521]: _type = "Task" [ 914.272390] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.280751] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f648ea-99d6-dd17-d198-2f32d17a04a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.292381] env[62521]: DEBUG nova.scheduler.client.report [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.522025] env[62521]: DEBUG nova.objects.base [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Object Instance<3805d897-fcb5-43ed-b502-7fad71124e74> lazy-loaded attributes: info_cache,migration_context {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 914.522696] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54db2d02-6af0-4371-beab-826e7181537c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.543250] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f0de652-251a-447d-bd55-c5394b4fccca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.549646] env[62521]: DEBUG oslo_vmware.api [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 914.549646] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fc125c-3a5f-7f38-241a-637c130364a7" [ 914.549646] env[62521]: _type = "Task" [ 914.549646] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.558535] env[62521]: DEBUG oslo_vmware.api [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fc125c-3a5f-7f38-241a-637c130364a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.589069] env[62521]: DEBUG oslo_vmware.api [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318755, 'name': ReconfigVM_Task, 'duration_secs': 0.423312} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.591619] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Reconfigured VM instance instance-00000037 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 914.597030] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-339db85e-36dc-410d-b5f0-77e40c3d85c4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.610616] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.614486] env[62521]: DEBUG oslo_vmware.api [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 914.614486] env[62521]: value = "task-1318757" [ 914.614486] env[62521]: _type = "Task" [ 914.614486] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.623298] env[62521]: DEBUG oslo_vmware.api [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318757, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.682131] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318756, 'name': Rename_Task, 'duration_secs': 0.443054} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.682423] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.682688] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7950f075-4283-423f-afe4-47ee009cdb26 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.688992] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 914.688992] env[62521]: value = "task-1318758" [ 914.688992] env[62521]: _type = "Task" [ 914.688992] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.697376] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318758, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.717021] env[62521]: DEBUG oslo_concurrency.lockutils [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] Releasing lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.717021] env[62521]: DEBUG nova.compute.manager [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Received event network-changed-9e954d34-4424-459c-a022-fc039668c616 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.717021] env[62521]: DEBUG nova.compute.manager [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Refreshing instance network info cache due to event network-changed-9e954d34-4424-459c-a022-fc039668c616. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 914.717021] env[62521]: DEBUG oslo_concurrency.lockutils [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] Acquiring lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.717512] env[62521]: DEBUG oslo_concurrency.lockutils [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] Acquired lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.717512] env[62521]: DEBUG nova.network.neutron [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Refreshing network info cache for port 9e954d34-4424-459c-a022-fc039668c616 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 914.782971] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f648ea-99d6-dd17-d198-2f32d17a04a8, 'name': SearchDatastore_Task, 'duration_secs': 0.027815} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.783383] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.783714] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.784053] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.784345] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.784594] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.784921] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c22a58dd-e63f-4f99-9c9a-337f973f37ac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.799132] env[62521]: DEBUG oslo_concurrency.lockutils [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.912s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.802174] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.802506] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.807262] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.112s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.809581] env[62521]: INFO nova.compute.claims [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.813544] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89d68e9a-77bb-4209-bda6-112eb577fed5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.823339] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 914.823339] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5206b78b-7626-6969-44c1-be71926f38ee" [ 914.823339] env[62521]: _type = "Task" [ 914.823339] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.827988] env[62521]: INFO nova.scheduler.client.report [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Deleted allocations for instance 6c4c36b7-b242-4578-9c23-d2529f308cb1 [ 914.832752] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5206b78b-7626-6969-44c1-be71926f38ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.987805] env[62521]: DEBUG nova.network.neutron [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating instance_info_cache with network_info: [{"id": "72671c1a-eabd-4a80-a591-17b26cea0249", "address": "fa:16:3e:3a:1f:96", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72671c1a-ea", "ovs_interfaceid": "72671c1a-eabd-4a80-a591-17b26cea0249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.061185] env[62521]: DEBUG oslo_vmware.api [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fc125c-3a5f-7f38-241a-637c130364a7, 'name': SearchDatastore_Task, 'duration_secs': 0.014253} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.061500] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.124592] env[62521]: DEBUG oslo_vmware.api [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318757, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.198589] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318758, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.334786] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5206b78b-7626-6969-44c1-be71926f38ee, 'name': SearchDatastore_Task, 'duration_secs': 0.023336} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.338994] env[62521]: DEBUG oslo_concurrency.lockutils [None req-285e04e0-1fa7-4e63-8550-2255b2e4552a tempest-ServersTestMultiNic-1072224182 tempest-ServersTestMultiNic-1072224182-project-member] Lock "6c4c36b7-b242-4578-9c23-d2529f308cb1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.737s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.343546] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcf09763-b961-4db0-b7e2-9c53c39e8896 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.350607] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 915.350607] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e939ad-a834-09dc-f043-d06c734bb49f" [ 915.350607] env[62521]: _type = "Task" [ 915.350607] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.363062] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e939ad-a834-09dc-f043-d06c734bb49f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.491055] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.630021] env[62521]: DEBUG oslo_vmware.api [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318757, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.704039] env[62521]: DEBUG oslo_vmware.api [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318758, 'name': PowerOnVM_Task, 'duration_secs': 0.847899} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.704124] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.704380] env[62521]: INFO nova.compute.manager [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Took 10.93 seconds to spawn the instance on the hypervisor. [ 915.704584] env[62521]: DEBUG nova.compute.manager [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.705416] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b21ab21-ee01-41b3-9762-cabeeadf0e38 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.788329] env[62521]: DEBUG nova.network.neutron [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updated VIF entry in instance network info cache for port 9e954d34-4424-459c-a022-fc039668c616. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 915.788700] env[62521]: DEBUG nova.network.neutron [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance_info_cache with network_info: [{"id": "9e954d34-4424-459c-a022-fc039668c616", "address": "fa:16:3e:90:66:a0", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e954d34-44", "ovs_interfaceid": "9e954d34-4424-459c-a022-fc039668c616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.867270] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e939ad-a834-09dc-f043-d06c734bb49f, 'name': SearchDatastore_Task, 'duration_secs': 0.015807} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.870638] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.871250] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 17876bcc-3a7b-47b0-afd9-1215e24864ae/17876bcc-3a7b-47b0-afd9-1215e24864ae.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 915.871852] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7ab3d1e-60d3-41f0-ba0e-ca4d773d33a6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.878925] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 915.878925] env[62521]: value = "task-1318759" [ 915.878925] env[62521]: _type = "Task" [ 915.878925] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.890572] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318759, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.070018] env[62521]: DEBUG nova.compute.manager [req-9a8894ca-7dbe-4c17-b8de-ce812d432726 req-9404bbbf-2ca1-415f-aa43-228935bda999 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received event network-vif-unplugged-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.070018] env[62521]: DEBUG oslo_concurrency.lockutils [req-9a8894ca-7dbe-4c17-b8de-ce812d432726 req-9404bbbf-2ca1-415f-aa43-228935bda999 service nova] Acquiring lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.070018] env[62521]: DEBUG oslo_concurrency.lockutils [req-9a8894ca-7dbe-4c17-b8de-ce812d432726 req-9404bbbf-2ca1-415f-aa43-228935bda999 service nova] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.070018] env[62521]: DEBUG oslo_concurrency.lockutils [req-9a8894ca-7dbe-4c17-b8de-ce812d432726 req-9404bbbf-2ca1-415f-aa43-228935bda999 service nova] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.070018] env[62521]: DEBUG nova.compute.manager [req-9a8894ca-7dbe-4c17-b8de-ce812d432726 req-9404bbbf-2ca1-415f-aa43-228935bda999 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] No waiting events found dispatching network-vif-unplugged-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 916.070018] env[62521]: WARNING nova.compute.manager [req-9a8894ca-7dbe-4c17-b8de-ce812d432726 req-9404bbbf-2ca1-415f-aa43-228935bda999 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received unexpected event network-vif-unplugged-72671c1a-eabd-4a80-a591-17b26cea0249 for instance with vm_state shelved and task_state shelving_offloading. [ 916.088021] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353f3ebb-afa6-4e40-b8ac-1576eabe34bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.095445] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5773d528-a513-48c1-b6a4-c551dd9c3690 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.134874] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622ff446-5b8f-45ff-acf3-32f56f3789d2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.146647] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac00b08c-1c1c-41c7-aeff-827d473050be {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.151548] env[62521]: DEBUG oslo_vmware.api [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318757, 'name': ReconfigVM_Task, 'duration_secs': 1.20742} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.152108] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282105', 'volume_id': '6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'name': 'volume-6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '538ed18e-920e-4ed4-a9e1-a6bfbae7f282', 'attached_at': '', 'detached_at': '', 'volume_id': '6e8eee93-7abc-4c71-8667-16b6b4dbd033', 'serial': '6e8eee93-7abc-4c71-8667-16b6b4dbd033'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 916.167248] env[62521]: DEBUG nova.compute.provider_tree [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.226797] env[62521]: INFO nova.compute.manager [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Took 30.47 seconds to build instance. [ 916.238552] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.240249] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d9b58e-b6d2-48b2-8c91-412d7d114c2d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.256157] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.256554] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5867a53b-305b-4107-ad85-6d3325fa1f0b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.294904] env[62521]: DEBUG oslo_concurrency.lockutils [req-fc6fad0d-7cc0-44e0-af56-a4c388640953 req-2d62e56b-ff7e-4917-8eda-f8ce06c280c4 service nova] Releasing lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.338052] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.338052] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.338052] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleting the datastore file [datastore2] 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.338052] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d57024b-942b-42e4-adfa-1f543d33b5e6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.344632] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 916.344632] env[62521]: value = "task-1318761" [ 916.344632] env[62521]: _type = "Task" [ 916.344632] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.353746] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318761, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.389927] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318759, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.670583] env[62521]: DEBUG nova.scheduler.client.report [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.710642] env[62521]: DEBUG nova.objects.instance [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'flavor' on Instance uuid 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.729600] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47ddc885-7cd1-4bbb-8005-caf05ed6bb34 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "611462d2-7f57-44b0-ac36-db32af3d0dd0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.988s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.862438] env[62521]: DEBUG oslo_vmware.api [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318761, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.302762} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.862438] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.862438] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 916.862438] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 916.903352] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318759, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595325} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.907571] env[62521]: INFO nova.scheduler.client.report [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleted allocations for instance 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad [ 916.908968] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 17876bcc-3a7b-47b0-afd9-1215e24864ae/17876bcc-3a7b-47b0-afd9-1215e24864ae.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 916.909227] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.909779] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7f84c68-4656-48ce-ba49-2b2713f4979a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.919313] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 916.919313] env[62521]: value = "task-1318762" [ 916.919313] env[62521]: _type = "Task" [ 916.919313] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.929392] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318762, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.178176] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.178176] env[62521]: DEBUG nova.compute.manager [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.183613] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.324s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.183613] env[62521]: DEBUG nova.objects.instance [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lazy-loading 'resources' on Instance uuid 4dcd1c6c-8726-42db-997a-e78ba1293310 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.397524] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.397903] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.398133] env[62521]: INFO nova.compute.manager [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Shelving [ 917.415862] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.429887] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318762, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075674} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.430182] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.434048] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef53f818-a36d-44aa-8da0-f94c2b8e7a37 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.460344] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 17876bcc-3a7b-47b0-afd9-1215e24864ae/17876bcc-3a7b-47b0-afd9-1215e24864ae.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.461832] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed8edc42-dd61-48fc-8fb3-bae0bf873410 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.482790] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "611462d2-7f57-44b0-ac36-db32af3d0dd0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.482978] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "611462d2-7f57-44b0-ac36-db32af3d0dd0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.483247] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "611462d2-7f57-44b0-ac36-db32af3d0dd0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.483474] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "611462d2-7f57-44b0-ac36-db32af3d0dd0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.483657] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "611462d2-7f57-44b0-ac36-db32af3d0dd0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.487269] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 917.487269] env[62521]: value = "task-1318763" [ 917.487269] env[62521]: _type = "Task" [ 917.487269] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.487767] env[62521]: INFO nova.compute.manager [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Terminating instance [ 917.492828] env[62521]: DEBUG nova.compute.manager [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 917.493042] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 917.494225] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11399a4-7fb8-47b0-983a-b44c423b597a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.502792] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318763, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.505947] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.505947] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a0c115a-eaff-4cac-88e3-f07532ec193e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.513086] env[62521]: DEBUG oslo_vmware.api [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 917.513086] env[62521]: value = "task-1318764" [ 917.513086] env[62521]: _type = "Task" [ 917.513086] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.527500] env[62521]: DEBUG oslo_vmware.api [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.687575] env[62521]: DEBUG nova.compute.utils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.691229] env[62521]: DEBUG nova.compute.manager [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 917.691569] env[62521]: DEBUG nova.network.neutron [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 917.720845] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c8978f36-931c-4753-b7c9-86bff4686f0e tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.324s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.786325] env[62521]: DEBUG nova.policy [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0e99fc3ead94078b185bc599ff247ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25420ed4f82e478989a011d40f037a24', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 917.790280] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.790490] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.790671] env[62521]: DEBUG nova.compute.manager [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.791933] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7af0878-7455-4618-8c0f-863f0dc91aa3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.803182] env[62521]: DEBUG nova.compute.manager [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62521) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 917.803807] env[62521]: DEBUG nova.objects.instance [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'flavor' on Instance uuid 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.905424] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.907013] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f31f081-e97f-49f5-8f4e-66465fa9ac49 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.908862] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-facbe5a7-dd51-4b95-aeba-2c6e4aea2f9c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.916271] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd292c8d-f63b-4e36-98b6-4545408dd4fe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.920861] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 917.920861] env[62521]: value = "task-1318765" [ 917.920861] env[62521]: _type = "Task" [ 917.920861] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.953375] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2475b458-6a45-4190-a151-57f3f8f9afb5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.960404] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318765, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.965840] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a612dc24-42a1-4bda-b314-2bb47a9fb016 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.982789] env[62521]: DEBUG nova.compute.provider_tree [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.999423] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318763, 'name': ReconfigVM_Task, 'duration_secs': 0.447942} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.999925] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 17876bcc-3a7b-47b0-afd9-1215e24864ae/17876bcc-3a7b-47b0-afd9-1215e24864ae.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.001080] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b33b803-504c-4565-b4ef-2abdf2ec35de {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.008394] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 918.008394] env[62521]: value = "task-1318766" [ 918.008394] env[62521]: _type = "Task" [ 918.008394] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.020867] env[62521]: DEBUG oslo_vmware.api [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318764, 'name': PowerOffVM_Task, 'duration_secs': 0.2041} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.025118] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.025118] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.025211] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318766, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.025395] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d6670a4-5d8d-4bdd-8baf-6ba491b3588d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.095454] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 918.095635] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 918.095877] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleting the datastore file [datastore1] 611462d2-7f57-44b0-ac36-db32af3d0dd0 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.096196] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94af7a24-92ea-4c20-8965-44ce66236aca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.107273] env[62521]: DEBUG oslo_vmware.api [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 918.107273] env[62521]: value = "task-1318768" [ 918.107273] env[62521]: _type = "Task" [ 918.107273] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.115773] env[62521]: DEBUG oslo_vmware.api [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318768, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.126580] env[62521]: DEBUG nova.compute.manager [req-d5155275-490a-485a-bb17-7cd18efd8e04 req-2a0a4601-14c0-4e5f-bb8d-319af99eecfb service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received event network-changed-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.126580] env[62521]: DEBUG nova.compute.manager [req-d5155275-490a-485a-bb17-7cd18efd8e04 req-2a0a4601-14c0-4e5f-bb8d-319af99eecfb service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Refreshing instance network info cache due to event network-changed-72671c1a-eabd-4a80-a591-17b26cea0249. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.126776] env[62521]: DEBUG oslo_concurrency.lockutils [req-d5155275-490a-485a-bb17-7cd18efd8e04 req-2a0a4601-14c0-4e5f-bb8d-319af99eecfb service nova] Acquiring lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.126776] env[62521]: DEBUG oslo_concurrency.lockutils [req-d5155275-490a-485a-bb17-7cd18efd8e04 req-2a0a4601-14c0-4e5f-bb8d-319af99eecfb service nova] Acquired lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.127318] env[62521]: DEBUG nova.network.neutron [req-d5155275-490a-485a-bb17-7cd18efd8e04 req-2a0a4601-14c0-4e5f-bb8d-319af99eecfb service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Refreshing network info cache for port 72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.192086] env[62521]: DEBUG nova.compute.manager [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.232287] env[62521]: DEBUG nova.network.neutron [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Successfully created port: 64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.314548] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 918.314912] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cfc43158-7485-4cb9-a2dd-beff6a7fd36f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.324694] env[62521]: DEBUG oslo_vmware.api [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 918.324694] env[62521]: value = "task-1318769" [ 918.324694] env[62521]: _type = "Task" [ 918.324694] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.337806] env[62521]: DEBUG oslo_vmware.api [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318769, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.430607] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318765, 'name': PowerOffVM_Task, 'duration_secs': 0.198049} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.430916] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.431716] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf36857-22ca-4b18-8056-87c88f26f181 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.451465] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7355b050-0f71-423a-861a-a0237334efab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.485833] env[62521]: DEBUG nova.scheduler.client.report [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.518341] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318766, 'name': Rename_Task, 'duration_secs': 0.151951} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.518863] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 918.519188] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53fe11f3-c062-4ab5-af91-856923ad3809 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.526984] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 918.526984] env[62521]: value = "task-1318770" [ 918.526984] env[62521]: _type = "Task" [ 918.526984] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.536095] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318770, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.615315] env[62521]: DEBUG oslo_vmware.api [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318768, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245866} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.615691] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.616492] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.616492] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.616492] env[62521]: INFO nova.compute.manager [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 918.616739] env[62521]: DEBUG oslo.service.loopingcall [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.616992] env[62521]: DEBUG nova.compute.manager [-] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 918.617149] env[62521]: DEBUG nova.network.neutron [-] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 918.836303] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.836303] env[62521]: DEBUG oslo_vmware.api [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318769, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.885912] env[62521]: DEBUG nova.network.neutron [req-d5155275-490a-485a-bb17-7cd18efd8e04 req-2a0a4601-14c0-4e5f-bb8d-319af99eecfb service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updated VIF entry in instance network info cache for port 72671c1a-eabd-4a80-a591-17b26cea0249. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 918.886266] env[62521]: DEBUG nova.network.neutron [req-d5155275-490a-485a-bb17-7cd18efd8e04 req-2a0a4601-14c0-4e5f-bb8d-319af99eecfb service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating instance_info_cache with network_info: [{"id": "72671c1a-eabd-4a80-a591-17b26cea0249", "address": "fa:16:3e:3a:1f:96", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap72671c1a-ea", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.962508] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 918.963368] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d0335fc6-af27-47fc-9538-4dceddbd4849 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.971697] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 918.971697] env[62521]: value = "task-1318771" [ 918.971697] env[62521]: _type = "Task" [ 918.971697] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.981275] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318771, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.991333] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.812s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.994678] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.383s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.995718] env[62521]: INFO nova.compute.claims [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.032065] env[62521]: INFO nova.scheduler.client.report [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleted allocations for instance 4dcd1c6c-8726-42db-997a-e78ba1293310 [ 919.047738] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318770, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.202471] env[62521]: DEBUG nova.compute.manager [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.238898] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.238898] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.239158] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.239310] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.239469] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.239619] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.239850] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.240043] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.241562] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.241766] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.241961] env[62521]: DEBUG nova.virt.hardware [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.242914] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72e5ba8-ee47-4fd3-936b-e79c7d88d342 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.255860] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81664a1f-00ad-4d04-af63-0996e156cfe0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.337746] env[62521]: DEBUG oslo_vmware.api [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318769, 'name': PowerOffVM_Task, 'duration_secs': 0.981698} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.337746] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 919.337746] env[62521]: DEBUG nova.compute.manager [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.338515] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425f9ae8-9a31-4613-ac41-776a427ece34 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.389495] env[62521]: DEBUG oslo_concurrency.lockutils [req-d5155275-490a-485a-bb17-7cd18efd8e04 req-2a0a4601-14c0-4e5f-bb8d-319af99eecfb service nova] Releasing lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.482231] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318771, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.537688] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318770, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.553490] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c6ab3275-734a-4cf2-ae7e-b2d8262f69b8 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "4dcd1c6c-8726-42db-997a-e78ba1293310" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.867s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.586753] env[62521]: DEBUG nova.network.neutron [-] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.852499] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e423bff4-b473-4ff7-8de8-7ef000e86d06 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.062s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.984020] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318771, 'name': CreateSnapshot_Task, 'duration_secs': 0.663632} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.984303] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 919.985039] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1342d2-75a7-4a65-9ff3-c6b596075344 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.046034] env[62521]: DEBUG oslo_vmware.api [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318770, 'name': PowerOnVM_Task, 'duration_secs': 1.076781} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.046034] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.046158] env[62521]: INFO nova.compute.manager [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Took 9.88 seconds to spawn the instance on the hypervisor. [ 920.046438] env[62521]: DEBUG nova.compute.manager [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.049599] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2824009a-25b3-4b13-a5f5-08bed8966ac8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.088717] env[62521]: INFO nova.compute.manager [-] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Took 1.47 seconds to deallocate network for instance. [ 920.276976] env[62521]: DEBUG nova.compute.manager [req-ae72402f-a3d2-4dfa-bbd4-642c1d730bf8 req-b9a357c7-b3d9-47e6-8800-a856baa76226 service nova] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Received event network-vif-deleted-4667150a-b284-4514-a637-498fbe7b40b3 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.296891] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cef6c14-45b5-4231-9687-388d50c45b04 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.309018] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed8c701-bfdc-40b5-b7ac-480288257bb7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.336587] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5138c5c3-aea6-4920-a861-e1814f6cd6eb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.343890] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8610de-9e84-4a77-9a7e-371b840096d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.357477] env[62521]: DEBUG nova.compute.provider_tree [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.521682] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 920.522012] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5e2decf0-4871-4583-ba14-d9e95b5ff84b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.530691] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 920.530691] env[62521]: value = "task-1318772" [ 920.530691] env[62521]: _type = "Task" [ 920.530691] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.540326] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318772, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.553596] env[62521]: DEBUG nova.objects.instance [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'flavor' on Instance uuid 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.569505] env[62521]: INFO nova.compute.manager [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Took 27.37 seconds to build instance. [ 920.610711] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.669590] env[62521]: DEBUG nova.network.neutron [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Successfully updated port: 64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 920.861072] env[62521]: DEBUG nova.scheduler.client.report [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.041596] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318772, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.059447] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.059741] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquired lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.060010] env[62521]: DEBUG nova.network.neutron [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 921.060340] env[62521]: DEBUG nova.objects.instance [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'info_cache' on Instance uuid 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.071847] env[62521]: DEBUG oslo_concurrency.lockutils [None req-215614b8-697f-476a-8f9b-b4be6d8e1806 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "17876bcc-3a7b-47b0-afd9-1215e24864ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.900s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.172479] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "refresh_cache-3afba9cd-f66e-44c5-b660-b4314f370901" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.172650] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "refresh_cache-3afba9cd-f66e-44c5-b660-b4314f370901" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.172756] env[62521]: DEBUG nova.network.neutron [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 921.366322] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.366923] env[62521]: DEBUG nova.compute.manager [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 921.369553] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 6.308s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.541767] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318772, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.563824] env[62521]: DEBUG nova.objects.base [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Object Instance<538ed18e-920e-4ed4-a9e1-a6bfbae7f282> lazy-loaded attributes: flavor,info_cache {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 921.719544] env[62521]: DEBUG nova.network.neutron [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.863685] env[62521]: DEBUG nova.network.neutron [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Updating instance_info_cache with network_info: [{"id": "64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee", "address": "fa:16:3e:ab:73:78", "network": {"id": "ce7ccdc7-e99d-48bd-93a9-c8c14725d58b", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1698306226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25420ed4f82e478989a011d40f037a24", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b56036cd-97ac-47f5-9089-7b38bfe99228", "external-id": "nsx-vlan-transportzone-301", "segmentation_id": 301, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64c6f406-e4", "ovs_interfaceid": "64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.872647] env[62521]: DEBUG nova.compute.utils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.877150] env[62521]: DEBUG nova.compute.manager [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 921.877327] env[62521]: DEBUG nova.network.neutron [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 921.916598] env[62521]: DEBUG nova.policy [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c63ec03cec5437285d6a99fc2a1d791', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '840dfca13260497b93990e4875cfd532', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.002571] env[62521]: DEBUG nova.compute.manager [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Stashing vm_state: active {{(pid=62521) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 922.046727] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318772, 'name': CloneVM_Task} progress is 95%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.078331] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a64ffc-534f-4383-8954-b910f2a86092 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.087067] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf4ee43-93e2-40e9-817b-26a39192f322 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.118743] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6323e04-8edc-4a87-8beb-7b510706a817 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.128969] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a648441-6100-444a-a1a2-3f09bebdb97b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.143995] env[62521]: DEBUG nova.compute.provider_tree [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.321201] env[62521]: DEBUG nova.compute.manager [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Received event network-vif-plugged-64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.321451] env[62521]: DEBUG oslo_concurrency.lockutils [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] Acquiring lock "3afba9cd-f66e-44c5-b660-b4314f370901-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.321693] env[62521]: DEBUG oslo_concurrency.lockutils [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] Lock "3afba9cd-f66e-44c5-b660-b4314f370901-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.321866] env[62521]: DEBUG oslo_concurrency.lockutils [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] Lock "3afba9cd-f66e-44c5-b660-b4314f370901-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.322050] env[62521]: DEBUG nova.compute.manager [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] No waiting events found dispatching network-vif-plugged-64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 922.322242] env[62521]: WARNING nova.compute.manager [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Received unexpected event network-vif-plugged-64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee for instance with vm_state building and task_state spawning. [ 922.322618] env[62521]: DEBUG nova.compute.manager [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Received event network-changed-64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.322618] env[62521]: DEBUG nova.compute.manager [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Refreshing instance network info cache due to event network-changed-64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 922.322740] env[62521]: DEBUG oslo_concurrency.lockutils [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] Acquiring lock "refresh_cache-3afba9cd-f66e-44c5-b660-b4314f370901" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.366858] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "refresh_cache-3afba9cd-f66e-44c5-b660-b4314f370901" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.367033] env[62521]: DEBUG nova.compute.manager [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Instance network_info: |[{"id": "64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee", "address": "fa:16:3e:ab:73:78", "network": {"id": "ce7ccdc7-e99d-48bd-93a9-c8c14725d58b", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1698306226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25420ed4f82e478989a011d40f037a24", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b56036cd-97ac-47f5-9089-7b38bfe99228", "external-id": "nsx-vlan-transportzone-301", "segmentation_id": 301, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64c6f406-e4", "ovs_interfaceid": "64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.367430] env[62521]: DEBUG oslo_concurrency.lockutils [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] Acquired lock "refresh_cache-3afba9cd-f66e-44c5-b660-b4314f370901" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.367682] env[62521]: DEBUG nova.network.neutron [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Refreshing network info cache for port 64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 922.369189] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:73:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b56036cd-97ac-47f5-9089-7b38bfe99228', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.378048] env[62521]: DEBUG oslo.service.loopingcall [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.379388] env[62521]: DEBUG nova.compute.manager [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 922.382061] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.382793] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c0b376a-682f-485d-bce9-5ab001568410 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.403662] env[62521]: DEBUG nova.network.neutron [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Successfully created port: f8c8e380-9d98-41ae-a53a-ba729b908294 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.406884] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.406884] env[62521]: value = "task-1318773" [ 922.406884] env[62521]: _type = "Task" [ 922.406884] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.414577] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318773, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.523029] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.525867] env[62521]: DEBUG nova.network.neutron [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updating instance_info_cache with network_info: [{"id": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "address": "fa:16:3e:41:9c:b1", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06aa91ed-3b", "ovs_interfaceid": "06aa91ed-3b1d-4c27-a91f-6246e75add4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.543312] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318772, 'name': CloneVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.647212] env[62521]: DEBUG nova.scheduler.client.report [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.919885] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318773, 'name': CreateVM_Task, 'duration_secs': 0.410126} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.922476] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 922.923234] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.923411] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.923744] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.924013] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a866071-d64b-43d4-9ebd-5a3b81756606 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.933209] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 922.933209] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]528fa077-2917-b61c-7578-40e7d6f4dcf5" [ 922.933209] env[62521]: _type = "Task" [ 922.933209] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.950692] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]528fa077-2917-b61c-7578-40e7d6f4dcf5, 'name': SearchDatastore_Task, 'duration_secs': 0.01125} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.950988] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.951242] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.951493] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.951651] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.951835] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.952106] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c4d8b9a-a1ce-44b7-bdca-9329d614857d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.961548] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.961738] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 922.962452] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6566c406-325f-46a9-a8b4-ce84dbe2a5fe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.968323] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 922.968323] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e2325d-28f9-1caa-1cc8-f9dadcfacb90" [ 922.968323] env[62521]: _type = "Task" [ 922.968323] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.976668] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e2325d-28f9-1caa-1cc8-f9dadcfacb90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.028421] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Releasing lock "refresh_cache-538ed18e-920e-4ed4-a9e1-a6bfbae7f282" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.049959] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318772, 'name': CloneVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.251351] env[62521]: DEBUG nova.network.neutron [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Updated VIF entry in instance network info cache for port 64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 923.251731] env[62521]: DEBUG nova.network.neutron [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Updating instance_info_cache with network_info: [{"id": "64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee", "address": "fa:16:3e:ab:73:78", "network": {"id": "ce7ccdc7-e99d-48bd-93a9-c8c14725d58b", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1698306226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25420ed4f82e478989a011d40f037a24", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b56036cd-97ac-47f5-9089-7b38bfe99228", "external-id": "nsx-vlan-transportzone-301", "segmentation_id": 301, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64c6f406-e4", "ovs_interfaceid": "64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.393725] env[62521]: DEBUG nova.compute.manager [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 923.419337] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.419599] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.419768] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.419956] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.420129] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.420285] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.420498] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.420664] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.420837] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.421012] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.421198] env[62521]: DEBUG nova.virt.hardware [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.422061] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d490d7fc-7dd7-4809-88fb-27dd514094f0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.430092] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82535cc4-c6e6-4510-9131-9bb5ca53e1ec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.476948] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e2325d-28f9-1caa-1cc8-f9dadcfacb90, 'name': SearchDatastore_Task, 'duration_secs': 0.008239} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.477737] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b6b7de8-b001-46f6-9bdd-27c9c860f5e1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.482602] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 923.482602] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5283e76d-5b7d-aca0-f16a-a39560712857" [ 923.482602] env[62521]: _type = "Task" [ 923.482602] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.491225] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5283e76d-5b7d-aca0-f16a-a39560712857, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.531982] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.532235] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a9e82b30-6037-40fe-b7be-4a27b43fb16d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.538554] env[62521]: DEBUG oslo_vmware.api [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 923.538554] env[62521]: value = "task-1318774" [ 923.538554] env[62521]: _type = "Task" [ 923.538554] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.549523] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318772, 'name': CloneVM_Task, 'duration_secs': 3.000584} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.552524] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Created linked-clone VM from snapshot [ 923.552780] env[62521]: DEBUG oslo_vmware.api [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.553467] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34794d65-fa8b-49d0-9c22-60dc1425efe2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.560688] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Uploading image 2919854a-4a07-48e6-87af-29f13a5c2313 {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 923.581938] env[62521]: DEBUG oslo_vmware.rw_handles [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 923.581938] env[62521]: value = "vm-282132" [ 923.581938] env[62521]: _type = "VirtualMachine" [ 923.581938] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 923.582194] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8a8e849a-31db-40ee-819c-db9989f8bcf6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.589012] env[62521]: DEBUG oslo_vmware.rw_handles [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lease: (returnval){ [ 923.589012] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fab843-67e6-edf5-47ae-a0a9cf263fa1" [ 923.589012] env[62521]: _type = "HttpNfcLease" [ 923.589012] env[62521]: } obtained for exporting VM: (result){ [ 923.589012] env[62521]: value = "vm-282132" [ 923.589012] env[62521]: _type = "VirtualMachine" [ 923.589012] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 923.589298] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the lease: (returnval){ [ 923.589298] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fab843-67e6-edf5-47ae-a0a9cf263fa1" [ 923.589298] env[62521]: _type = "HttpNfcLease" [ 923.589298] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 923.595233] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 923.595233] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fab843-67e6-edf5-47ae-a0a9cf263fa1" [ 923.595233] env[62521]: _type = "HttpNfcLease" [ 923.595233] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 923.663913] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.294s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.666877] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.252s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.667205] env[62521]: DEBUG nova.objects.instance [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'resources' on Instance uuid 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.754422] env[62521]: DEBUG oslo_concurrency.lockutils [req-eda84cab-2155-41da-ac06-7a9ec2aaa6c2 req-a4ba4a67-df1b-40d9-9594-569b115ba784 service nova] Releasing lock "refresh_cache-3afba9cd-f66e-44c5-b660-b4314f370901" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.993938] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5283e76d-5b7d-aca0-f16a-a39560712857, 'name': SearchDatastore_Task, 'duration_secs': 0.030493} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.994240] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.994539] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 3afba9cd-f66e-44c5-b660-b4314f370901/3afba9cd-f66e-44c5-b660-b4314f370901.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 923.994820] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c5a5dd0-3044-4db2-b939-710e382e5d35 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.001222] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 924.001222] env[62521]: value = "task-1318776" [ 924.001222] env[62521]: _type = "Task" [ 924.001222] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.009123] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318776, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.048547] env[62521]: DEBUG oslo_vmware.api [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318774, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.097959] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 924.097959] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fab843-67e6-edf5-47ae-a0a9cf263fa1" [ 924.097959] env[62521]: _type = "HttpNfcLease" [ 924.097959] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 924.098553] env[62521]: DEBUG oslo_vmware.rw_handles [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 924.098553] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fab843-67e6-edf5-47ae-a0a9cf263fa1" [ 924.098553] env[62521]: _type = "HttpNfcLease" [ 924.098553] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 924.099508] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc3f0d1-d4b6-4a8d-b3fb-7bebdc880e48 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.107818] env[62521]: DEBUG oslo_vmware.rw_handles [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5251d8ec-0855-ce91-6d3a-f1bef83746db/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 924.108462] env[62521]: DEBUG oslo_vmware.rw_handles [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5251d8ec-0855-ce91-6d3a-f1bef83746db/disk-0.vmdk for reading. {{(pid=62521) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 924.179788] env[62521]: DEBUG nova.objects.instance [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'numa_topology' on Instance uuid 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.218243] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b6b4a15c-9c08-4fee-aa60-2ce610e1ac5d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.250593] env[62521]: INFO nova.scheduler.client.report [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted allocation for migration 225a66f1-5d99-45a5-a517-4041e11d90f1 [ 924.358481] env[62521]: DEBUG nova.compute.manager [req-199f98e1-4551-4a24-8b09-99c021cf19fa req-33d8df0c-0df9-446c-b24f-dabd291be006 service nova] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Received event network-vif-plugged-f8c8e380-9d98-41ae-a53a-ba729b908294 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.358760] env[62521]: DEBUG oslo_concurrency.lockutils [req-199f98e1-4551-4a24-8b09-99c021cf19fa req-33d8df0c-0df9-446c-b24f-dabd291be006 service nova] Acquiring lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.359019] env[62521]: DEBUG oslo_concurrency.lockutils [req-199f98e1-4551-4a24-8b09-99c021cf19fa req-33d8df0c-0df9-446c-b24f-dabd291be006 service nova] Lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.359561] env[62521]: DEBUG oslo_concurrency.lockutils [req-199f98e1-4551-4a24-8b09-99c021cf19fa req-33d8df0c-0df9-446c-b24f-dabd291be006 service nova] Lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.359799] env[62521]: DEBUG nova.compute.manager [req-199f98e1-4551-4a24-8b09-99c021cf19fa req-33d8df0c-0df9-446c-b24f-dabd291be006 service nova] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] No waiting events found dispatching network-vif-plugged-f8c8e380-9d98-41ae-a53a-ba729b908294 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 924.360068] env[62521]: WARNING nova.compute.manager [req-199f98e1-4551-4a24-8b09-99c021cf19fa req-33d8df0c-0df9-446c-b24f-dabd291be006 service nova] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Received unexpected event network-vif-plugged-f8c8e380-9d98-41ae-a53a-ba729b908294 for instance with vm_state building and task_state spawning. [ 924.468513] env[62521]: DEBUG nova.network.neutron [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Successfully updated port: f8c8e380-9d98-41ae-a53a-ba729b908294 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.512406] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318776, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.549332] env[62521]: DEBUG oslo_vmware.api [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318774, 'name': PowerOnVM_Task, 'duration_secs': 0.582433} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.549709] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.549955] env[62521]: DEBUG nova.compute.manager [None req-bf55c5d6-47e0-4cd0-a8b6-db57f72a7818 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.550913] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67f3274-02a1-45f7-8dd3-3df8f781a543 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.686053] env[62521]: DEBUG nova.objects.base [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Object Instance<73bc1cf1-7ee1-4a9b-8214-2dc752b700ad> lazy-loaded attributes: resources,numa_topology {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 924.761412] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8288b5fc-2e8b-438b-af19-6b0e11e663dc tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "3805d897-fcb5-43ed-b502-7fad71124e74" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.328s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.937103] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b9fd8a-2f45-4782-b627-a15dedb35220 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.945031] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae0faf3-9047-4115-a8f1-c132f4bb1f67 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.976123] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "refresh_cache-cc4b29ee-969d-49c6-9fae-71d5f850aae1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.976414] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired lock "refresh_cache-cc4b29ee-969d-49c6-9fae-71d5f850aae1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.976723] env[62521]: DEBUG nova.network.neutron [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.979226] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28be8e11-78c7-43ac-adc4-00f05009a31a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.989272] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c29652e6-29b8-4720-b9bc-e3bdc5a26782 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.011193] env[62521]: DEBUG nova.compute.provider_tree [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.023447] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318776, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555755} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.023736] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 3afba9cd-f66e-44c5-b660-b4314f370901/3afba9cd-f66e-44c5-b660-b4314f370901.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.024140] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.024790] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a65a1724-9164-4457-b9ef-b99ce653f092 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.032013] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 925.032013] env[62521]: value = "task-1318777" [ 925.032013] env[62521]: _type = "Task" [ 925.032013] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.044083] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318777, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.518048] env[62521]: DEBUG nova.scheduler.client.report [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.534529] env[62521]: DEBUG nova.network.neutron [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.549159] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318777, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07008} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.549460] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.550412] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d62be7-58b1-4be1-b3ec-d90bdb674139 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.574353] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 3afba9cd-f66e-44c5-b660-b4314f370901/3afba9cd-f66e-44c5-b660-b4314f370901.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.577245] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef50f833-9967-43ba-b654-9fa027e0998d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.598693] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 925.598693] env[62521]: value = "task-1318778" [ 925.598693] env[62521]: _type = "Task" [ 925.598693] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.606823] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318778, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.735612] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "3805d897-fcb5-43ed-b502-7fad71124e74" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.735896] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "3805d897-fcb5-43ed-b502-7fad71124e74" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.736174] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "3805d897-fcb5-43ed-b502-7fad71124e74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.736555] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "3805d897-fcb5-43ed-b502-7fad71124e74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.736998] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "3805d897-fcb5-43ed-b502-7fad71124e74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.739503] env[62521]: INFO nova.compute.manager [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Terminating instance [ 925.742499] env[62521]: DEBUG nova.compute.manager [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 925.742724] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.743575] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f370f9cd-bb51-49b2-a712-20ef9c92cc6e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.752071] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.752834] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84a6a1b0-b640-47f3-9e90-69bb93646b8b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.759228] env[62521]: DEBUG oslo_vmware.api [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 925.759228] env[62521]: value = "task-1318779" [ 925.759228] env[62521]: _type = "Task" [ 925.759228] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.767652] env[62521]: DEBUG oslo_vmware.api [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318779, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.972051] env[62521]: DEBUG nova.network.neutron [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Updating instance_info_cache with network_info: [{"id": "f8c8e380-9d98-41ae-a53a-ba729b908294", "address": "fa:16:3e:26:14:28", "network": {"id": "5b74eda2-668c-4d8f-b4d8-a5e748613ee1", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-574755447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "840dfca13260497b93990e4875cfd532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8c8e380-9d", "ovs_interfaceid": "f8c8e380-9d98-41ae-a53a-ba729b908294", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.023471] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.356s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.026398] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.417s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.026775] env[62521]: DEBUG nova.objects.instance [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lazy-loading 'resources' on Instance uuid 611462d2-7f57-44b0-ac36-db32af3d0dd0 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.115268] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318778, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.269698] env[62521]: DEBUG oslo_vmware.api [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318779, 'name': PowerOffVM_Task, 'duration_secs': 0.293668} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.269698] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.269698] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.270099] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55fd68c8-e94f-447d-bdcf-837be729562a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.338940] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 926.339324] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 926.339599] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleting the datastore file [datastore2] 3805d897-fcb5-43ed-b502-7fad71124e74 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 926.340100] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-253e4ac2-5e9a-4e92-9053-25495415c665 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.346667] env[62521]: DEBUG oslo_vmware.api [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 926.346667] env[62521]: value = "task-1318781" [ 926.346667] env[62521]: _type = "Task" [ 926.346667] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.356856] env[62521]: DEBUG oslo_vmware.api [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318781, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.394783] env[62521]: DEBUG nova.compute.manager [req-f86f3fba-9ae1-4d1c-b4b1-9f7bbd267bc8 req-0305ff64-471c-4a8e-8907-fe5a4060a95d service nova] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Received event network-changed-f8c8e380-9d98-41ae-a53a-ba729b908294 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.395017] env[62521]: DEBUG nova.compute.manager [req-f86f3fba-9ae1-4d1c-b4b1-9f7bbd267bc8 req-0305ff64-471c-4a8e-8907-fe5a4060a95d service nova] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Refreshing instance network info cache due to event network-changed-f8c8e380-9d98-41ae-a53a-ba729b908294. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 926.395371] env[62521]: DEBUG oslo_concurrency.lockutils [req-f86f3fba-9ae1-4d1c-b4b1-9f7bbd267bc8 req-0305ff64-471c-4a8e-8907-fe5a4060a95d service nova] Acquiring lock "refresh_cache-cc4b29ee-969d-49c6-9fae-71d5f850aae1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.473906] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Releasing lock "refresh_cache-cc4b29ee-969d-49c6-9fae-71d5f850aae1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.474310] env[62521]: DEBUG nova.compute.manager [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Instance network_info: |[{"id": "f8c8e380-9d98-41ae-a53a-ba729b908294", "address": "fa:16:3e:26:14:28", "network": {"id": "5b74eda2-668c-4d8f-b4d8-a5e748613ee1", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-574755447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "840dfca13260497b93990e4875cfd532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8c8e380-9d", "ovs_interfaceid": "f8c8e380-9d98-41ae-a53a-ba729b908294", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 926.474688] env[62521]: DEBUG oslo_concurrency.lockutils [req-f86f3fba-9ae1-4d1c-b4b1-9f7bbd267bc8 req-0305ff64-471c-4a8e-8907-fe5a4060a95d service nova] Acquired lock "refresh_cache-cc4b29ee-969d-49c6-9fae-71d5f850aae1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.474882] env[62521]: DEBUG nova.network.neutron [req-f86f3fba-9ae1-4d1c-b4b1-9f7bbd267bc8 req-0305ff64-471c-4a8e-8907-fe5a4060a95d service nova] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Refreshing network info cache for port f8c8e380-9d98-41ae-a53a-ba729b908294 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 926.476246] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:14:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '415e68b4-3766-4359-afe2-f8563910d98c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8c8e380-9d98-41ae-a53a-ba729b908294', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.484533] env[62521]: DEBUG oslo.service.loopingcall [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.485732] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.485975] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89dd9d8a-6b0c-4644-acdb-869ea667f421 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.506381] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.506381] env[62521]: value = "task-1318782" [ 926.506381] env[62521]: _type = "Task" [ 926.506381] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.514201] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318782, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.535189] env[62521]: DEBUG oslo_concurrency.lockutils [None req-21d09e57-3f4a-4f35-a015-ce6c7fe08061 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 30.723s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.536545] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 7.701s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.536778] env[62521]: INFO nova.compute.manager [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Unshelving [ 926.553629] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.553987] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.609193] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318778, 'name': ReconfigVM_Task, 'duration_secs': 0.54626} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.611902] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 3afba9cd-f66e-44c5-b660-b4314f370901/3afba9cd-f66e-44c5-b660-b4314f370901.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.612753] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b6e379f-3ec9-434d-8c77-42d4d852eea1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.619239] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 926.619239] env[62521]: value = "task-1318783" [ 926.619239] env[62521]: _type = "Task" [ 926.619239] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.630754] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318783, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.764627] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ee85ac-6503-4997-a084-eabe0fb47736 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.773167] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa59b1ef-7803-4118-9fbb-2a06402229ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.810099] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b86613-6b33-4e68-a1af-9226071b8e12 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.818061] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89fff72-d060-49fd-a556-9c0b9c8b1edb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.833266] env[62521]: DEBUG nova.compute.provider_tree [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.858053] env[62521]: DEBUG oslo_vmware.api [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318781, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152842} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.858331] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.858555] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 926.858829] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 926.859073] env[62521]: INFO nova.compute.manager [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Took 1.12 seconds to destroy the instance on the hypervisor. [ 926.859498] env[62521]: DEBUG oslo.service.loopingcall [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.859574] env[62521]: DEBUG nova.compute.manager [-] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 926.859694] env[62521]: DEBUG nova.network.neutron [-] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 927.016638] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318782, 'name': CreateVM_Task, 'duration_secs': 0.476921} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.016853] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.017791] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.018015] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.018378] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.018686] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-304e1b05-688f-4fd8-80ff-97eeca4119c4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.023992] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 927.023992] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52747da2-12e5-ccac-d32f-e86007e3ce2a" [ 927.023992] env[62521]: _type = "Task" [ 927.023992] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.033459] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52747da2-12e5-ccac-d32f-e86007e3ce2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.058066] env[62521]: DEBUG nova.compute.manager [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 927.129854] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318783, 'name': Rename_Task, 'duration_secs': 0.15456} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.130168] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.130426] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d29d2036-4981-48ab-896d-81ab9373eaf2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.136607] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 927.136607] env[62521]: value = "task-1318784" [ 927.136607] env[62521]: _type = "Task" [ 927.136607] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.149022] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318784, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.337380] env[62521]: DEBUG nova.scheduler.client.report [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.458269] env[62521]: DEBUG nova.network.neutron [req-f86f3fba-9ae1-4d1c-b4b1-9f7bbd267bc8 req-0305ff64-471c-4a8e-8907-fe5a4060a95d service nova] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Updated VIF entry in instance network info cache for port f8c8e380-9d98-41ae-a53a-ba729b908294. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 927.458564] env[62521]: DEBUG nova.network.neutron [req-f86f3fba-9ae1-4d1c-b4b1-9f7bbd267bc8 req-0305ff64-471c-4a8e-8907-fe5a4060a95d service nova] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Updating instance_info_cache with network_info: [{"id": "f8c8e380-9d98-41ae-a53a-ba729b908294", "address": "fa:16:3e:26:14:28", "network": {"id": "5b74eda2-668c-4d8f-b4d8-a5e748613ee1", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-574755447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "840dfca13260497b93990e4875cfd532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "415e68b4-3766-4359-afe2-f8563910d98c", "external-id": "nsx-vlan-transportzone-538", "segmentation_id": 538, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8c8e380-9d", "ovs_interfaceid": "f8c8e380-9d98-41ae-a53a-ba729b908294", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.535750] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52747da2-12e5-ccac-d32f-e86007e3ce2a, 'name': SearchDatastore_Task, 'duration_secs': 0.011644} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.536164] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.536480] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.536631] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.536724] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.536919] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.537229] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3deada6-5758-46ed-adad-6b0cb3f8e351 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.547082] env[62521]: DEBUG nova.compute.utils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 927.548440] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.548623] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 927.549629] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8aa0c4df-3ea6-495a-89fc-b8afebe5fcb9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.554911] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 927.554911] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522eb334-a41c-64a3-be3d-a97c9c38a284" [ 927.554911] env[62521]: _type = "Task" [ 927.554911] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.565815] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522eb334-a41c-64a3-be3d-a97c9c38a284, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.592420] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.649106] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318784, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.846216] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.820s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.849188] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.326s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.859788] env[62521]: DEBUG nova.network.neutron [-] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.867686] env[62521]: INFO nova.scheduler.client.report [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted allocations for instance 611462d2-7f57-44b0-ac36-db32af3d0dd0 [ 927.961749] env[62521]: DEBUG oslo_concurrency.lockutils [req-f86f3fba-9ae1-4d1c-b4b1-9f7bbd267bc8 req-0305ff64-471c-4a8e-8907-fe5a4060a95d service nova] Releasing lock "refresh_cache-cc4b29ee-969d-49c6-9fae-71d5f850aae1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.051729] env[62521]: INFO nova.virt.block_device [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Booting with volume a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c at /dev/sdb [ 928.069734] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522eb334-a41c-64a3-be3d-a97c9c38a284, 'name': SearchDatastore_Task, 'duration_secs': 0.008363} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.070562] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1267026d-3d84-49ac-9a65-6e249ecf6182 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.076591] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 928.076591] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521125b5-48ae-b56b-8339-7baf9cb6e6f4" [ 928.076591] env[62521]: _type = "Task" [ 928.076591] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.085575] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521125b5-48ae-b56b-8339-7baf9cb6e6f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.097680] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77642a31-f595-40b2-a59b-190eb4853457 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.106229] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef6165e-12b8-4ffb-9fe5-d8ba0d9c89ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.136578] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0331c3b0-9543-4615-8953-e70243567b5d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.147579] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318784, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.151501] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7c3045-bb64-4471-95f6-5514fced7efa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.188107] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7d3c31-9431-4231-9a41-f1f7c6067f34 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.197795] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52c3330-f714-476f-871a-9c2016d5dba7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.213498] env[62521]: DEBUG nova.virt.block_device [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating existing volume attachment record: ba14b520-08f7-464f-8613-1c2608bdfe0f {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 928.353591] env[62521]: INFO nova.compute.claims [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.362130] env[62521]: INFO nova.compute.manager [-] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Took 1.50 seconds to deallocate network for instance. [ 928.375419] env[62521]: DEBUG oslo_concurrency.lockutils [None req-29a0d832-f8e1-4409-b36e-3c94ee8d390b tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "611462d2-7f57-44b0-ac36-db32af3d0dd0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.892s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.429970] env[62521]: DEBUG nova.compute.manager [req-eedd81d0-6ccc-41a8-a3f7-a18adcaaa4de req-8541896b-f76a-4153-8ccb-798254dfc2c9 service nova] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Received event network-vif-deleted-1643da59-fa10-4645-994b-0c557c3a6e10 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.588566] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521125b5-48ae-b56b-8339-7baf9cb6e6f4, 'name': SearchDatastore_Task, 'duration_secs': 0.015102} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.589472] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.589849] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] cc4b29ee-969d-49c6-9fae-71d5f850aae1/cc4b29ee-969d-49c6-9fae-71d5f850aae1.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 928.590291] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2fcd9c9-a16a-4b81-8274-43a8fdfd0ab5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.600127] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 928.600127] env[62521]: value = "task-1318788" [ 928.600127] env[62521]: _type = "Task" [ 928.600127] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.610581] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.649025] env[62521]: DEBUG oslo_vmware.api [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318784, 'name': PowerOnVM_Task, 'duration_secs': 1.034012} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.649279] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.649381] env[62521]: INFO nova.compute.manager [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Took 9.45 seconds to spawn the instance on the hypervisor. [ 928.649568] env[62521]: DEBUG nova.compute.manager [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.650409] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1166aa6e-3373-4d98-bdba-c4f3490ad981 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.861876] env[62521]: INFO nova.compute.resource_tracker [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating resource usage from migration a367e195-f862-4960-ab2d-b0ece3adece2 [ 928.868459] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.111516] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318788, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.113548] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b88f1a-2eb8-4b96-916b-dcaf2b460775 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.122635] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a441613-5472-4689-88eb-59808ab481ec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.156044] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80b078a-3eca-43b1-a75c-fe20e983e869 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.170119] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb3879d-88a6-4482-90aa-d334bff6a5b4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.174843] env[62521]: INFO nova.compute.manager [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Took 17.49 seconds to build instance. [ 929.194131] env[62521]: DEBUG nova.compute.provider_tree [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.611298] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.666957} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.611579] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] cc4b29ee-969d-49c6-9fae-71d5f850aae1/cc4b29ee-969d-49c6-9fae-71d5f850aae1.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.611715] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.612040] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-573577ed-cfa9-4898-a7e9-a12b42dfce6e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.619349] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 929.619349] env[62521]: value = "task-1318789" [ 929.619349] env[62521]: _type = "Task" [ 929.619349] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.630059] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318789, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.634016] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "ac420a77-2160-45d6-b313-c5304536c39d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.634327] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "ac420a77-2160-45d6-b313-c5304536c39d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.678750] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5481820a-5827-4440-8b2a-7fd27f1d056b tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "3afba9cd-f66e-44c5-b660-b4314f370901" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.007s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.696546] env[62521]: DEBUG nova.scheduler.client.report [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.130292] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318789, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084489} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.130590] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.131552] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00dc499-fccd-42bb-aa83-cf4c157e2834 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.145269] env[62521]: DEBUG nova.compute.manager [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 930.156181] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] cc4b29ee-969d-49c6-9fae-71d5f850aae1/cc4b29ee-969d-49c6-9fae-71d5f850aae1.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.156720] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f954df28-95c2-4ae1-a705-c0b2e15f0c70 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.176673] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 930.176673] env[62521]: value = "task-1318790" [ 930.176673] env[62521]: _type = "Task" [ 930.176673] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.188775] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318790, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.202312] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.354s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.202494] env[62521]: INFO nova.compute.manager [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Migrating [ 930.208905] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.617s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.210299] env[62521]: INFO nova.compute.claims [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.678725] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.688398] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318790, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.721960] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.722150] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.722339] env[62521]: DEBUG nova.network.neutron [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.859196] env[62521]: DEBUG nova.compute.manager [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.860218] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9143d5c-9703-4636-8ca7-de5f615c08c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.187545] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318790, 'name': ReconfigVM_Task, 'duration_secs': 0.680792} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.187888] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Reconfigured VM instance instance-00000051 to attach disk [datastore2] cc4b29ee-969d-49c6-9fae-71d5f850aae1/cc4b29ee-969d-49c6-9fae-71d5f850aae1.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.188617] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b000a86-7a76-45d3-b86b-9e52401526fb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.195067] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 931.195067] env[62521]: value = "task-1318792" [ 931.195067] env[62521]: _type = "Task" [ 931.195067] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.203981] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318792, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.373344] env[62521]: INFO nova.compute.manager [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] instance snapshotting [ 931.379129] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3c4a33-b4a0-4734-ab24-00023c060d6b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.405267] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44dd7490-f424-46e9-8515-464b22f85697 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.483662] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622c15ab-87bc-4299-b960-1e902b40a0ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.487396] env[62521]: DEBUG nova.network.neutron [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance_info_cache with network_info: [{"id": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "address": "fa:16:3e:1a:80:af", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0934da-f4", "ovs_interfaceid": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.493573] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed53206-48d2-4614-82bd-490ff33f5c76 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.526897] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e3c076-0706-431c-9ec6-8e645d373e83 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.534463] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d06620f-f1a7-4a06-82e3-3363b5a2fa52 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.547989] env[62521]: DEBUG nova.compute.provider_tree [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.706497] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318792, 'name': Rename_Task, 'duration_secs': 0.223691} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.706949] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.707073] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61d87c09-0c24-4153-9f3f-e5bef8e293e1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.713470] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 931.713470] env[62521]: value = "task-1318793" [ 931.713470] env[62521]: _type = "Task" [ 931.713470] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.721045] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318793, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.919495] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 931.919854] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6cecf381-f166-4c56-b32d-bf83114db5f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.928709] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 931.928709] env[62521]: value = "task-1318794" [ 931.928709] env[62521]: _type = "Task" [ 931.928709] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.938292] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318794, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.991226] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.051020] env[62521]: DEBUG nova.scheduler.client.report [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.223663] env[62521]: DEBUG oslo_vmware.api [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318793, 'name': PowerOnVM_Task, 'duration_secs': 0.474893} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.223956] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.224185] env[62521]: INFO nova.compute.manager [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Took 8.83 seconds to spawn the instance on the hypervisor. [ 932.224403] env[62521]: DEBUG nova.compute.manager [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.225223] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57bc18b9-8033-433b-b413-9882c981df86 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.440439] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318794, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.484478] env[62521]: DEBUG oslo_vmware.rw_handles [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5251d8ec-0855-ce91-6d3a-f1bef83746db/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 932.485410] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5291cf12-5641-45f7-9fbd-6a42066253fc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.491367] env[62521]: DEBUG oslo_vmware.rw_handles [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5251d8ec-0855-ce91-6d3a-f1bef83746db/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 932.491542] env[62521]: ERROR oslo_vmware.rw_handles [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5251d8ec-0855-ce91-6d3a-f1bef83746db/disk-0.vmdk due to incomplete transfer. [ 932.491760] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-37ef72c3-c8f7-42ac-a9b9-9eefe250f2a3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.501533] env[62521]: DEBUG oslo_vmware.rw_handles [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5251d8ec-0855-ce91-6d3a-f1bef83746db/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 932.501700] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Uploaded image 2919854a-4a07-48e6-87af-29f13a5c2313 to the Glance image server {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 932.503801] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 932.504042] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-cf6b9821-a790-46d4-b293-e195228dd5e5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.509498] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 932.509498] env[62521]: value = "task-1318795" [ 932.509498] env[62521]: _type = "Task" [ 932.509498] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.517617] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318795, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.555397] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.346s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.555932] env[62521]: DEBUG nova.compute.manager [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.558566] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.690s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.558763] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.560784] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.882s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.562683] env[62521]: INFO nova.compute.claims [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 932.582248] env[62521]: INFO nova.scheduler.client.report [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted allocations for instance 3805d897-fcb5-43ed-b502-7fad71124e74 [ 932.743023] env[62521]: INFO nova.compute.manager [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Took 18.16 seconds to build instance. [ 932.941205] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318794, 'name': CreateSnapshot_Task, 'duration_secs': 0.993781} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.941503] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 932.942306] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c275aa9f-d577-4f72-860c-b1c31ccf774d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.019323] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318795, 'name': Destroy_Task, 'duration_secs': 0.439601} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.019602] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Destroyed the VM [ 933.019843] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 933.020120] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-84edc104-ef0e-48f1-8d4d-88045c8dabd5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.026463] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 933.026463] env[62521]: value = "task-1318796" [ 933.026463] env[62521]: _type = "Task" [ 933.026463] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.035290] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318796, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.067565] env[62521]: DEBUG nova.compute.utils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.071321] env[62521]: DEBUG nova.compute.manager [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.071517] env[62521]: DEBUG nova.network.neutron [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 933.095874] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6c4efbcc-f4af-4532-a5a0-a83c926845d3 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "3805d897-fcb5-43ed-b502-7fad71124e74" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.360s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.172409] env[62521]: DEBUG nova.policy [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d7aab6f3f6245c2b56979677428a9df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e6baab1556e4b00b07ab554664b4445', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 933.245453] env[62521]: DEBUG oslo_concurrency.lockutils [None req-023fc300-8819-4207-b611-58458a6c7d64 tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.667s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.461050] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 933.461050] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8c85b05a-6c3d-431b-86f2-c46da5e01769 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.469869] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 933.469869] env[62521]: value = "task-1318797" [ 933.469869] env[62521]: _type = "Task" [ 933.469869] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.477572] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318797, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.504512] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74298e9b-9e20-4c6e-8194-78c9e5bea9e5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.524145] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance '17876bcc-3a7b-47b0-afd9-1215e24864ae' progress to 0 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 933.535999] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318796, 'name': RemoveSnapshot_Task, 'duration_secs': 0.374837} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.539849] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 933.539849] env[62521]: DEBUG nova.compute.manager [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.541312] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fd11a5-f23c-46c3-a305-a8577ad14330 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.572905] env[62521]: DEBUG nova.compute.manager [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.594835] env[62521]: DEBUG oslo_concurrency.lockutils [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.596054] env[62521]: DEBUG oslo_concurrency.lockutils [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.596295] env[62521]: DEBUG oslo_concurrency.lockutils [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.596654] env[62521]: DEBUG oslo_concurrency.lockutils [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.597738] env[62521]: DEBUG oslo_concurrency.lockutils [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.599406] env[62521]: INFO nova.compute.manager [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Terminating instance [ 933.602947] env[62521]: DEBUG nova.network.neutron [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Successfully created port: 34cee820-50a0-47bd-a0c7-25f1e01fa639 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.606181] env[62521]: DEBUG nova.compute.manager [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 933.606323] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.607639] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4aeed21-a1fe-443c-a78f-27535fd81c54 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.618093] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.618369] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e43ac6d6-7f15-477f-8881-77b03172bd92 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.627395] env[62521]: DEBUG oslo_vmware.api [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 933.627395] env[62521]: value = "task-1318798" [ 933.627395] env[62521]: _type = "Task" [ 933.627395] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.638532] env[62521]: DEBUG oslo_vmware.api [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318798, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.807794] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03104ce-c785-4ccb-a059-f19728fe8f4f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.815753] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef7fe4c-8673-47e7-a955-2002fb070f76 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.850376] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b1395e-7fa5-44f6-8965-957a0baa1fba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.858579] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3263c175-e63d-4061-908a-6e9a28fe2208 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.863830] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.877377] env[62521]: DEBUG nova.compute.provider_tree [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.979922] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318797, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.033617] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.034078] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d74fef63-0af9-4a3c-93a2-933f613d2938 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.041794] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 934.041794] env[62521]: value = "task-1318799" [ 934.041794] env[62521]: _type = "Task" [ 934.041794] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.050664] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318799, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.054892] env[62521]: INFO nova.compute.manager [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Shelve offloading [ 934.060048] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.061271] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49ec63d3-7c73-4eff-b383-52ecb4c18561 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.068421] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 934.068421] env[62521]: value = "task-1318800" [ 934.068421] env[62521]: _type = "Task" [ 934.068421] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.079146] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 934.079531] env[62521]: DEBUG nova.compute.manager [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.080508] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f5bd00-12d0-471a-a808-e883265a0063 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.090711] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.090711] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.090934] env[62521]: DEBUG nova.network.neutron [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.139997] env[62521]: DEBUG oslo_vmware.api [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318798, 'name': PowerOffVM_Task, 'duration_secs': 0.279399} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.140507] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.140608] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.140821] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6082f901-92fc-476d-b49e-c32acc2d409c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.380715] env[62521]: DEBUG nova.scheduler.client.report [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.480899] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318797, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.551541] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318799, 'name': PowerOffVM_Task, 'duration_secs': 0.195703} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.551891] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.552110] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance '17876bcc-3a7b-47b0-afd9-1215e24864ae' progress to 17 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 934.589655] env[62521]: DEBUG nova.compute.manager [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.622753] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.622753] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.622753] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.623017] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.623017] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.623134] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.623347] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.623530] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.623703] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.623891] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.624066] env[62521]: DEBUG nova.virt.hardware [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.625314] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f4a69f-d4dd-47fc-8b06-ec26c05d744f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.633862] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397ef92a-b970-4303-8197-71ceb476a509 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.776989] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 934.776989] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 934.777267] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Deleting the datastore file [datastore2] cc4b29ee-969d-49c6-9fae-71d5f850aae1 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.777573] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41c511c8-e49e-4929-96fe-1eb77df44710 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.784410] env[62521]: DEBUG oslo_vmware.api [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for the task: (returnval){ [ 934.784410] env[62521]: value = "task-1318802" [ 934.784410] env[62521]: _type = "Task" [ 934.784410] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.793457] env[62521]: DEBUG oslo_vmware.api [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318802, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.857713] env[62521]: DEBUG nova.network.neutron [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updating instance_info_cache with network_info: [{"id": "903b74d0-11d2-404d-81f1-b6bb18f713af", "address": "fa:16:3e:6a:82:e1", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap903b74d0-11", "ovs_interfaceid": "903b74d0-11d2-404d-81f1-b6bb18f713af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.887334] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.887334] env[62521]: DEBUG nova.compute.manager [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 934.889437] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.026s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.889676] env[62521]: DEBUG nova.objects.instance [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'pci_requests' on Instance uuid 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.981270] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318797, 'name': CloneVM_Task} progress is 95%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.031384] env[62521]: DEBUG nova.compute.manager [req-d2fe27ac-0589-43f9-9d3a-124ab6987c11 req-156b1d28-18c7-4958-88c5-3c248d828202 service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Received event network-vif-plugged-34cee820-50a0-47bd-a0c7-25f1e01fa639 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.031606] env[62521]: DEBUG oslo_concurrency.lockutils [req-d2fe27ac-0589-43f9-9d3a-124ab6987c11 req-156b1d28-18c7-4958-88c5-3c248d828202 service nova] Acquiring lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.031982] env[62521]: DEBUG oslo_concurrency.lockutils [req-d2fe27ac-0589-43f9-9d3a-124ab6987c11 req-156b1d28-18c7-4958-88c5-3c248d828202 service nova] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.032199] env[62521]: DEBUG oslo_concurrency.lockutils [req-d2fe27ac-0589-43f9-9d3a-124ab6987c11 req-156b1d28-18c7-4958-88c5-3c248d828202 service nova] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.032377] env[62521]: DEBUG nova.compute.manager [req-d2fe27ac-0589-43f9-9d3a-124ab6987c11 req-156b1d28-18c7-4958-88c5-3c248d828202 service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] No waiting events found dispatching network-vif-plugged-34cee820-50a0-47bd-a0c7-25f1e01fa639 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.032546] env[62521]: WARNING nova.compute.manager [req-d2fe27ac-0589-43f9-9d3a-124ab6987c11 req-156b1d28-18c7-4958-88c5-3c248d828202 service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Received unexpected event network-vif-plugged-34cee820-50a0-47bd-a0c7-25f1e01fa639 for instance with vm_state building and task_state spawning. [ 935.058159] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.058423] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.058589] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.058840] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.059016] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.059177] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.059388] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.059551] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.059732] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.059888] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.060077] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.065635] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f25eecfb-6599-4df7-ba0b-493b42ee3c62 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.081588] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 935.081588] env[62521]: value = "task-1318803" [ 935.081588] env[62521]: _type = "Task" [ 935.081588] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.091294] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318803, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.123935] env[62521]: DEBUG nova.network.neutron [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Successfully updated port: 34cee820-50a0-47bd-a0c7-25f1e01fa639 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.294590] env[62521]: DEBUG oslo_vmware.api [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Task: {'id': task-1318802, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194292} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.294860] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.295059] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.295251] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.295425] env[62521]: INFO nova.compute.manager [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Took 1.69 seconds to destroy the instance on the hypervisor. [ 935.295666] env[62521]: DEBUG oslo.service.loopingcall [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.295855] env[62521]: DEBUG nova.compute.manager [-] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.295947] env[62521]: DEBUG nova.network.neutron [-] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 935.360954] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.392476] env[62521]: DEBUG nova.compute.utils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 935.395202] env[62521]: DEBUG nova.objects.instance [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'numa_topology' on Instance uuid 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.398569] env[62521]: DEBUG nova.compute.manager [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 935.398743] env[62521]: DEBUG nova.network.neutron [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 935.441395] env[62521]: DEBUG nova.policy [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76d85009c2c1416b9ea049003e6d96a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c533bd4917b4466e9c5c76a9caf8e807', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 935.482937] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318797, 'name': CloneVM_Task, 'duration_secs': 1.849695} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.483240] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Created linked-clone VM from snapshot [ 935.484165] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10a98c4-0cff-4bcd-aa52-dc1cd8cb0edc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.491809] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Uploading image 9169c65d-1ed0-4d12-bbfd-189634f114ff {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 935.502645] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 935.502971] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f9f62637-9f12-440c-aabb-f53cca9b12cb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.510015] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 935.510015] env[62521]: value = "task-1318804" [ 935.510015] env[62521]: _type = "Task" [ 935.510015] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.519497] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318804, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.592157] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318803, 'name': ReconfigVM_Task, 'duration_secs': 0.186578} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.592611] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance '17876bcc-3a7b-47b0-afd9-1215e24864ae' progress to 33 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 935.627201] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.627393] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.627524] env[62521]: DEBUG nova.network.neutron [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.824039] env[62521]: DEBUG nova.network.neutron [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Successfully created port: 00b03967-a3d2-4262-9736-6cef5c454664 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 935.899554] env[62521]: DEBUG nova.compute.manager [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 935.902970] env[62521]: INFO nova.compute.claims [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.945680] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 935.946679] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23af38a0-0415-40cc-8a3c-dd7dc37229e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.955406] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.955679] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ebafa1f9-8dba-4ea8-ba12-01792011602e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.018466] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 936.018742] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 936.019316] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleting the datastore file [datastore1] bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.019738] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc91240f-54c3-4bee-8d55-03072cc2737d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.025598] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318804, 'name': Destroy_Task, 'duration_secs': 0.369121} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.026295] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Destroyed the VM [ 936.026691] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 936.026980] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d8490c7c-cacf-4c28-920e-1cbe764d91b6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.030080] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 936.030080] env[62521]: value = "task-1318806" [ 936.030080] env[62521]: _type = "Task" [ 936.030080] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.034452] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 936.034452] env[62521]: value = "task-1318807" [ 936.034452] env[62521]: _type = "Task" [ 936.034452] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.041289] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.046156] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318807, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.099441] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:36:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='e986ebeb-2013-4e50-9994-d8d533ed7ed2',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-778228551',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.099725] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.099857] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.100227] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.100376] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.100558] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.100790] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.100949] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.101155] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.101487] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.101567] env[62521]: DEBUG nova.virt.hardware [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.107437] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Reconfiguring VM instance instance-0000004f to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 936.107770] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8fd8b8a8-9619-436e-beb2-d6ad543d195b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.126778] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 936.126778] env[62521]: value = "task-1318808" [ 936.126778] env[62521]: _type = "Task" [ 936.126778] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.137112] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318808, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.163161] env[62521]: DEBUG nova.network.neutron [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.285385] env[62521]: DEBUG nova.network.neutron [-] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.542725] env[62521]: DEBUG oslo_vmware.api [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318806, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188205} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.543318] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 936.543513] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 936.543690] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 936.548147] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318807, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.568695] env[62521]: INFO nova.scheduler.client.report [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleted allocations for instance bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 [ 936.579012] env[62521]: DEBUG nova.network.neutron [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance_info_cache with network_info: [{"id": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "address": "fa:16:3e:8c:86:c8", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34cee820-50", "ovs_interfaceid": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.641025] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318808, 'name': ReconfigVM_Task, 'duration_secs': 0.187128} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.641025] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Reconfigured VM instance instance-0000004f to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 936.641025] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2370cf1-7e6c-4319-944e-08bdaed66f7e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.664065] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 17876bcc-3a7b-47b0-afd9-1215e24864ae/17876bcc-3a7b-47b0-afd9-1215e24864ae.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.664588] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07fb04cc-931d-4feb-91bf-67e4d8a3607d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.686894] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 936.686894] env[62521]: value = "task-1318809" [ 936.686894] env[62521]: _type = "Task" [ 936.686894] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.694681] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318809, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.788674] env[62521]: INFO nova.compute.manager [-] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Took 1.49 seconds to deallocate network for instance. [ 936.918445] env[62521]: DEBUG nova.compute.manager [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 936.948420] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.949251] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.949251] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.949251] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.949251] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.949438] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.949606] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.949775] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.949999] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.950134] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.950316] env[62521]: DEBUG nova.virt.hardware [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.951229] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22f5895-edc8-40ba-b2da-ccba04751c17 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.962341] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed84338a-2481-4b94-bc66-0134b5faa35d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.045677] env[62521]: DEBUG oslo_vmware.api [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318807, 'name': RemoveSnapshot_Task, 'duration_secs': 0.773591} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.048334] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 937.072702] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.081595] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.081863] env[62521]: DEBUG nova.compute.manager [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Instance network_info: |[{"id": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "address": "fa:16:3e:8c:86:c8", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34cee820-50", "ovs_interfaceid": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.082522] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:86:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34cee820-50a0-47bd-a0c7-25f1e01fa639', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.090317] env[62521]: DEBUG oslo.service.loopingcall [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.093017] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 937.093546] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6173cd28-6d27-4688-aae8-2e1083add5b0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.117996] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.117996] env[62521]: value = "task-1318810" [ 937.117996] env[62521]: _type = "Task" [ 937.117996] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.126388] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0f16f6-15eb-4bc8-a484-1ee745e53040 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.132841] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318810, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.137495] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-851ea425-e6c8-4b3b-9b83-448c9e9d6e65 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.174771] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817ff43e-7aaf-4806-820c-feb6c3afc8fc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.182503] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1199d55b-0a5e-41b6-9cd4-a333803fc1d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.193848] env[62521]: DEBUG nova.compute.manager [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Received event network-changed-34cee820-50a0-47bd-a0c7-25f1e01fa639 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.194138] env[62521]: DEBUG nova.compute.manager [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Refreshing instance network info cache due to event network-changed-34cee820-50a0-47bd-a0c7-25f1e01fa639. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.194293] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] Acquiring lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.194471] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] Acquired lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.194647] env[62521]: DEBUG nova.network.neutron [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Refreshing network info cache for port 34cee820-50a0-47bd-a0c7-25f1e01fa639 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.207319] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318809, 'name': ReconfigVM_Task, 'duration_secs': 0.269049} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.208217] env[62521]: DEBUG nova.compute.provider_tree [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.209702] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 17876bcc-3a7b-47b0-afd9-1215e24864ae/17876bcc-3a7b-47b0-afd9-1215e24864ae.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.209994] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance '17876bcc-3a7b-47b0-afd9-1215e24864ae' progress to 50 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 937.295495] env[62521]: DEBUG oslo_concurrency.lockutils [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.312473] env[62521]: DEBUG nova.compute.manager [req-1d5d54ad-db59-4f6a-842e-b4b2cc57cad8 req-41ebdf78-8ef7-4aa0-99ac-ac43cd957c3b service nova] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Received event network-vif-plugged-00b03967-a3d2-4262-9736-6cef5c454664 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.312473] env[62521]: DEBUG oslo_concurrency.lockutils [req-1d5d54ad-db59-4f6a-842e-b4b2cc57cad8 req-41ebdf78-8ef7-4aa0-99ac-ac43cd957c3b service nova] Acquiring lock "ac420a77-2160-45d6-b313-c5304536c39d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.312473] env[62521]: DEBUG oslo_concurrency.lockutils [req-1d5d54ad-db59-4f6a-842e-b4b2cc57cad8 req-41ebdf78-8ef7-4aa0-99ac-ac43cd957c3b service nova] Lock "ac420a77-2160-45d6-b313-c5304536c39d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.312473] env[62521]: DEBUG oslo_concurrency.lockutils [req-1d5d54ad-db59-4f6a-842e-b4b2cc57cad8 req-41ebdf78-8ef7-4aa0-99ac-ac43cd957c3b service nova] Lock "ac420a77-2160-45d6-b313-c5304536c39d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.312766] env[62521]: DEBUG nova.compute.manager [req-1d5d54ad-db59-4f6a-842e-b4b2cc57cad8 req-41ebdf78-8ef7-4aa0-99ac-ac43cd957c3b service nova] [instance: ac420a77-2160-45d6-b313-c5304536c39d] No waiting events found dispatching network-vif-plugged-00b03967-a3d2-4262-9736-6cef5c454664 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.313112] env[62521]: WARNING nova.compute.manager [req-1d5d54ad-db59-4f6a-842e-b4b2cc57cad8 req-41ebdf78-8ef7-4aa0-99ac-ac43cd957c3b service nova] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Received unexpected event network-vif-plugged-00b03967-a3d2-4262-9736-6cef5c454664 for instance with vm_state building and task_state spawning. [ 937.552941] env[62521]: WARNING nova.compute.manager [None req-01f78890-d13a-467d-b557-cd3e435c1ed5 tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Image not found during snapshot: nova.exception.ImageNotFound: Image 9169c65d-1ed0-4d12-bbfd-189634f114ff could not be found. [ 937.627668] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318810, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.715150] env[62521]: DEBUG nova.scheduler.client.report [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.720880] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6c90a1-ed8d-4b67-b04d-58db39abb7c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.740592] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c63096f-3821-4dd2-b979-1782992625b6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.758685] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance '17876bcc-3a7b-47b0-afd9-1215e24864ae' progress to 67 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 937.881351] env[62521]: DEBUG nova.network.neutron [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Successfully updated port: 00b03967-a3d2-4262-9736-6cef5c454664 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.904063] env[62521]: DEBUG nova.compute.manager [req-d74fcf25-f316-40b4-bb3f-1d717e3a677e req-de712c50-c463-4296-b9b9-c8559be6350e service nova] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Received event network-changed-00b03967-a3d2-4262-9736-6cef5c454664 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.904063] env[62521]: DEBUG nova.compute.manager [req-d74fcf25-f316-40b4-bb3f-1d717e3a677e req-de712c50-c463-4296-b9b9-c8559be6350e service nova] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Refreshing instance network info cache due to event network-changed-00b03967-a3d2-4262-9736-6cef5c454664. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.904212] env[62521]: DEBUG oslo_concurrency.lockutils [req-d74fcf25-f316-40b4-bb3f-1d717e3a677e req-de712c50-c463-4296-b9b9-c8559be6350e service nova] Acquiring lock "refresh_cache-ac420a77-2160-45d6-b313-c5304536c39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.904310] env[62521]: DEBUG oslo_concurrency.lockutils [req-d74fcf25-f316-40b4-bb3f-1d717e3a677e req-de712c50-c463-4296-b9b9-c8559be6350e service nova] Acquired lock "refresh_cache-ac420a77-2160-45d6-b313-c5304536c39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.904500] env[62521]: DEBUG nova.network.neutron [req-d74fcf25-f316-40b4-bb3f-1d717e3a677e req-de712c50-c463-4296-b9b9-c8559be6350e service nova] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Refreshing network info cache for port 00b03967-a3d2-4262-9736-6cef5c454664 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.989495] env[62521]: DEBUG nova.network.neutron [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updated VIF entry in instance network info cache for port 34cee820-50a0-47bd-a0c7-25f1e01fa639. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 937.989884] env[62521]: DEBUG nova.network.neutron [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance_info_cache with network_info: [{"id": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "address": "fa:16:3e:8c:86:c8", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34cee820-50", "ovs_interfaceid": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.129150] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318810, 'name': CreateVM_Task, 'duration_secs': 0.609183} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.129334] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.130168] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.130286] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.130622] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.130906] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b86efb7-20ad-42c8-9fe8-04a91a27d0f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.136055] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 938.136055] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a36965-d6c0-dd18-e14c-ce78af1f0d82" [ 938.136055] env[62521]: _type = "Task" [ 938.136055] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.139477] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.144160] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a36965-d6c0-dd18-e14c-ce78af1f0d82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.221513] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.332s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.223683] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.151s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.223928] env[62521]: DEBUG nova.objects.instance [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lazy-loading 'resources' on Instance uuid bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.268844] env[62521]: INFO nova.network.neutron [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating port 72671c1a-eabd-4a80-a591-17b26cea0249 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 938.310978] env[62521]: DEBUG nova.network.neutron [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Port 3b0934da-f42d-4e91-ae7c-a4f254d6b86c binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 938.384395] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "refresh_cache-ac420a77-2160-45d6-b313-c5304536c39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.435943] env[62521]: DEBUG nova.network.neutron [req-d74fcf25-f316-40b4-bb3f-1d717e3a677e req-de712c50-c463-4296-b9b9-c8559be6350e service nova] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 938.496302] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] Releasing lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.496302] env[62521]: DEBUG nova.compute.manager [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Received event network-vif-deleted-f8c8e380-9d98-41ae-a53a-ba729b908294 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.496302] env[62521]: DEBUG nova.compute.manager [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Received event network-vif-unplugged-903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.496302] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] Acquiring lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.496302] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.496302] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.496302] env[62521]: DEBUG nova.compute.manager [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] No waiting events found dispatching network-vif-unplugged-903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 938.496302] env[62521]: WARNING nova.compute.manager [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Received unexpected event network-vif-unplugged-903b74d0-11d2-404d-81f1-b6bb18f713af for instance with vm_state shelved_offloaded and task_state unshelving. [ 938.496302] env[62521]: DEBUG nova.compute.manager [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Received event network-changed-903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.496302] env[62521]: DEBUG nova.compute.manager [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Refreshing instance network info cache due to event network-changed-903b74d0-11d2-404d-81f1-b6bb18f713af. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 938.496302] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] Acquiring lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.496302] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] Acquired lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.496302] env[62521]: DEBUG nova.network.neutron [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Refreshing network info cache for port 903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.508570] env[62521]: DEBUG nova.network.neutron [req-d74fcf25-f316-40b4-bb3f-1d717e3a677e req-de712c50-c463-4296-b9b9-c8559be6350e service nova] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.647950] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a36965-d6c0-dd18-e14c-ce78af1f0d82, 'name': SearchDatastore_Task, 'duration_secs': 0.009548} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.647950] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.647950] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.648137] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.648176] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.648347] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.648609] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-759ebc34-8aa3-4427-a70c-10352ec2ad59 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.657100] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.657282] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 938.657946] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88da4c6a-8a8d-49ea-84e9-257780f0167d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.662646] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 938.662646] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d29f94-f5e4-af24-8a7c-3d210c9e6f0d" [ 938.662646] env[62521]: _type = "Task" [ 938.662646] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.670409] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d29f94-f5e4-af24-8a7c-3d210c9e6f0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.726066] env[62521]: DEBUG nova.objects.instance [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lazy-loading 'numa_topology' on Instance uuid bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.850188] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "3afba9cd-f66e-44c5-b660-b4314f370901" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.850188] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "3afba9cd-f66e-44c5-b660-b4314f370901" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.850514] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "3afba9cd-f66e-44c5-b660-b4314f370901-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.850562] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "3afba9cd-f66e-44c5-b660-b4314f370901-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.850708] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "3afba9cd-f66e-44c5-b660-b4314f370901-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.853394] env[62521]: INFO nova.compute.manager [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Terminating instance [ 938.855294] env[62521]: DEBUG nova.compute.manager [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.855683] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.856377] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1b0baf-d552-4ca2-a8ec-c68b5d2cc993 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.864894] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.865157] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03da7139-44ad-4ffd-83a2-82fe41647752 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.871345] env[62521]: DEBUG oslo_vmware.api [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 938.871345] env[62521]: value = "task-1318811" [ 938.871345] env[62521]: _type = "Task" [ 938.871345] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.879322] env[62521]: DEBUG oslo_vmware.api [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318811, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.009999] env[62521]: DEBUG oslo_concurrency.lockutils [req-d74fcf25-f316-40b4-bb3f-1d717e3a677e req-de712c50-c463-4296-b9b9-c8559be6350e service nova] Releasing lock "refresh_cache-ac420a77-2160-45d6-b313-c5304536c39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.010399] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "refresh_cache-ac420a77-2160-45d6-b313-c5304536c39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.010560] env[62521]: DEBUG nova.network.neutron [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.172321] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d29f94-f5e4-af24-8a7c-3d210c9e6f0d, 'name': SearchDatastore_Task, 'duration_secs': 0.009622} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.173423] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4adbc8b-2471-485d-a901-2c0364cbc457 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.178230] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 939.178230] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525865a1-ad3c-571c-22db-e4ee83d1a073" [ 939.178230] env[62521]: _type = "Task" [ 939.178230] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.185558] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525865a1-ad3c-571c-22db-e4ee83d1a073, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.193505] env[62521]: DEBUG nova.network.neutron [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updated VIF entry in instance network info cache for port 903b74d0-11d2-404d-81f1-b6bb18f713af. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.193847] env[62521]: DEBUG nova.network.neutron [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updating instance_info_cache with network_info: [{"id": "903b74d0-11d2-404d-81f1-b6bb18f713af", "address": "fa:16:3e:6a:82:e1", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": null, "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap903b74d0-11", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.229017] env[62521]: DEBUG nova.objects.base [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 939.332990] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "17876bcc-3a7b-47b0-afd9-1215e24864ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.333249] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "17876bcc-3a7b-47b0-afd9-1215e24864ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.333432] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "17876bcc-3a7b-47b0-afd9-1215e24864ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.382781] env[62521]: DEBUG oslo_vmware.api [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318811, 'name': PowerOffVM_Task, 'duration_secs': 0.191332} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.383167] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.383249] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.383505] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a19195a4-fe4b-48cd-a9db-a7c45cee19fe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.417240] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c9de1d-9617-4629-a0ad-14f9d52b1025 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.424773] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bcb7b65-c736-4455-bdbf-62634e9fa8ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.453454] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb39e24-3d69-418b-ad66-8e4798bc20cf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.460845] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c58053e-ab34-4871-ae91-c97c19334ba0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.473609] env[62521]: DEBUG nova.compute.provider_tree [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.491339] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.491566] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.491756] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleting the datastore file [datastore2] 3afba9cd-f66e-44c5-b660-b4314f370901 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.492021] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-615bd2cf-292d-4e6f-b2d3-7dfcb9b5366a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.497930] env[62521]: DEBUG oslo_vmware.api [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for the task: (returnval){ [ 939.497930] env[62521]: value = "task-1318813" [ 939.497930] env[62521]: _type = "Task" [ 939.497930] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.505684] env[62521]: DEBUG oslo_vmware.api [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318813, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.544264] env[62521]: DEBUG nova.network.neutron [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 939.670350] env[62521]: DEBUG nova.network.neutron [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Updating instance_info_cache with network_info: [{"id": "00b03967-a3d2-4262-9736-6cef5c454664", "address": "fa:16:3e:9a:de:c2", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00b03967-a3", "ovs_interfaceid": "00b03967-a3d2-4262-9736-6cef5c454664", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.689267] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525865a1-ad3c-571c-22db-e4ee83d1a073, 'name': SearchDatastore_Task, 'duration_secs': 0.009741} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.689580] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.689854] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c/02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 939.690159] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c19ffd4-430c-4046-bbf9-7eda7fd4f32d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.699017] env[62521]: DEBUG oslo_concurrency.lockutils [req-7cce1aa8-2407-4ed8-9798-8c6e275f1485 req-904d37bd-27c7-422c-a522-04d2402a1d9c service nova] Releasing lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.699017] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 939.699017] env[62521]: value = "task-1318814" [ 939.699017] env[62521]: _type = "Task" [ 939.699017] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.706580] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318814, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.796611] env[62521]: DEBUG nova.compute.manager [req-a061ae63-aa89-45f7-aa99-1bb804e5bc92 req-0e1029b1-6f30-4a57-add6-f6401cc65193 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received event network-vif-plugged-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.796883] env[62521]: DEBUG oslo_concurrency.lockutils [req-a061ae63-aa89-45f7-aa99-1bb804e5bc92 req-0e1029b1-6f30-4a57-add6-f6401cc65193 service nova] Acquiring lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.797242] env[62521]: DEBUG oslo_concurrency.lockutils [req-a061ae63-aa89-45f7-aa99-1bb804e5bc92 req-0e1029b1-6f30-4a57-add6-f6401cc65193 service nova] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.797492] env[62521]: DEBUG oslo_concurrency.lockutils [req-a061ae63-aa89-45f7-aa99-1bb804e5bc92 req-0e1029b1-6f30-4a57-add6-f6401cc65193 service nova] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.797735] env[62521]: DEBUG nova.compute.manager [req-a061ae63-aa89-45f7-aa99-1bb804e5bc92 req-0e1029b1-6f30-4a57-add6-f6401cc65193 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] No waiting events found dispatching network-vif-plugged-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 939.797933] env[62521]: WARNING nova.compute.manager [req-a061ae63-aa89-45f7-aa99-1bb804e5bc92 req-0e1029b1-6f30-4a57-add6-f6401cc65193 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received unexpected event network-vif-plugged-72671c1a-eabd-4a80-a591-17b26cea0249 for instance with vm_state shelved_offloaded and task_state spawning. [ 939.948476] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.949342] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.949638] env[62521]: DEBUG nova.network.neutron [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.976979] env[62521]: DEBUG nova.scheduler.client.report [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.012225] env[62521]: DEBUG oslo_vmware.api [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Task: {'id': task-1318813, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152897} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.012570] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.012759] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.013704] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.013960] env[62521]: INFO nova.compute.manager [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Took 1.16 seconds to destroy the instance on the hypervisor. [ 940.014441] env[62521]: DEBUG oslo.service.loopingcall [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.014997] env[62521]: DEBUG nova.compute.manager [-] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.015148] env[62521]: DEBUG nova.network.neutron [-] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 940.173216] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "refresh_cache-ac420a77-2160-45d6-b313-c5304536c39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.173646] env[62521]: DEBUG nova.compute.manager [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Instance network_info: |[{"id": "00b03967-a3d2-4262-9736-6cef5c454664", "address": "fa:16:3e:9a:de:c2", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00b03967-a3", "ovs_interfaceid": "00b03967-a3d2-4262-9736-6cef5c454664", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 940.174218] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:de:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '84aee122-f630-43c5-9cc1-3a38d3819c82', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00b03967-a3d2-4262-9736-6cef5c454664', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.182580] env[62521]: DEBUG oslo.service.loopingcall [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.182834] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.183091] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc3bc6ae-0dd7-4d3b-98f4-8fd706ded63e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.206620] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318814, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.208381] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.208381] env[62521]: value = "task-1318815" [ 940.208381] env[62521]: _type = "Task" [ 940.208381] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.216401] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318815, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.368689] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.368896] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.369095] env[62521]: DEBUG nova.network.neutron [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.486989] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.263s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.489576] env[62521]: DEBUG oslo_concurrency.lockutils [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.194s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.489807] env[62521]: DEBUG nova.objects.instance [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lazy-loading 'resources' on Instance uuid cc4b29ee-969d-49c6-9fae-71d5f850aae1 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.714012] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318814, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527097} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.716775] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c/02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 940.717049] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.717328] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ddda5b21-73da-4f20-ac4a-5e3cb78e10f7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.724895] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318815, 'name': CreateVM_Task, 'duration_secs': 0.354053} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.726183] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 940.726813] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 940.726813] env[62521]: value = "task-1318816" [ 940.726813] env[62521]: _type = "Task" [ 940.726813] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.727245] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.727426] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.727790] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 940.728135] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-253ae6b2-dccd-4849-9c77-458c52bc0c1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.738697] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 940.738697] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d2ae00-2526-1cb1-d62c-1f00b10e71c1" [ 940.738697] env[62521]: _type = "Task" [ 940.738697] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.742309] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318816, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.751519] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d2ae00-2526-1cb1-d62c-1f00b10e71c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.816327] env[62521]: DEBUG nova.network.neutron [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating instance_info_cache with network_info: [{"id": "72671c1a-eabd-4a80-a591-17b26cea0249", "address": "fa:16:3e:3a:1f:96", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72671c1a-ea", "ovs_interfaceid": "72671c1a-eabd-4a80-a591-17b26cea0249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.827917] env[62521]: DEBUG nova.network.neutron [-] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.000341] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68528a93-a8d4-48a4-9501-6f017b13a7e2 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.602s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.001159] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.862s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.001352] env[62521]: INFO nova.compute.manager [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Unshelving [ 941.117210] env[62521]: DEBUG nova.network.neutron [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance_info_cache with network_info: [{"id": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "address": "fa:16:3e:1a:80:af", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0934da-f4", "ovs_interfaceid": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.225482] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ce5e04-661c-4924-aa15-0018cac367f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.237351] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318816, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084448} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.239311] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.240161] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5aee50-7655-4617-bf5f-aa9469756498 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.243157] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b6db2b-8f9a-402f-b713-824c68e9be58 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.254860] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d2ae00-2526-1cb1-d62c-1f00b10e71c1, 'name': SearchDatastore_Task, 'duration_secs': 0.011731} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.287740] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.288121] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.288456] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.288670] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.288904] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.298390] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c/02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.298711] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27a30b9d-a9de-4d3a-b0d7-d2003814fcd6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.301158] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0dec7e-27c2-4319-a9bd-f3e064ef4acb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.305224] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d862d46c-3555-42be-958b-46b9b531a27d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.319772] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.328769] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954d9a6f-ae5b-457e-9b57-5bcfc65774e0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.333053] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 941.333053] env[62521]: value = "task-1318817" [ 941.333053] env[62521]: _type = "Task" [ 941.333053] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.334238] env[62521]: INFO nova.compute.manager [-] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Took 1.32 seconds to deallocate network for instance. [ 941.334621] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.334802] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.339804] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-898577fe-717c-486b-a628-8f15c536e8d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.353970] env[62521]: DEBUG nova.compute.provider_tree [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.360677] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318817, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.363047] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='47fa5e64cdf6751606a577faab33e665',container_format='bare',created_at=2024-10-20T00:36:19Z,direct_url=,disk_format='vmdk',id=720eb7fe-5500-4cf9-a06e-3471ca9179ed,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-99101872-shelved',owner='6ab001785c25414abbaa4bfe3a6a0bb4',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-20T00:36:35Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.363222] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.363312] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.363500] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.363651] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.363801] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.364046] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.364194] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.364366] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.364576] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.364761] env[62521]: DEBUG nova.virt.hardware [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.366478] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a15b5b-ad0a-444a-9906-3e75e0260d8c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.369092] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 941.369092] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527b60f1-b516-fd6b-5cd9-a5ba81cd2942" [ 941.369092] env[62521]: _type = "Task" [ 941.369092] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.376817] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63596c88-d2cd-4cd2-9eb6-550fb85d16cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.384058] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527b60f1-b516-fd6b-5cd9-a5ba81cd2942, 'name': SearchDatastore_Task, 'duration_secs': 0.011938} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.385131] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-405a1eeb-d964-439a-b203-330e7da968b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.395470] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:1f:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ec3f9e71-839a-429d-b211-d3dfc98ca4f6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72671c1a-eabd-4a80-a591-17b26cea0249', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.402671] env[62521]: DEBUG oslo.service.loopingcall [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.403469] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 941.403707] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a37d1956-7a98-4119-97c9-0c53c204a065 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.419183] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 941.419183] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e566ff-d1cb-ecf6-6357-5887e9ad47a3" [ 941.419183] env[62521]: _type = "Task" [ 941.419183] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.424307] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.424307] env[62521]: value = "task-1318818" [ 941.424307] env[62521]: _type = "Task" [ 941.424307] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.427419] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e566ff-d1cb-ecf6-6357-5887e9ad47a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.434356] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318818, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.620073] env[62521]: DEBUG oslo_concurrency.lockutils [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.831277] env[62521]: DEBUG nova.compute.manager [req-b81858ae-e7d2-4fe9-82a0-f0c30bd3ba3b req-1a51c932-f914-4fef-8525-25ec1fedcc28 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received event network-changed-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.831490] env[62521]: DEBUG nova.compute.manager [req-b81858ae-e7d2-4fe9-82a0-f0c30bd3ba3b req-1a51c932-f914-4fef-8525-25ec1fedcc28 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Refreshing instance network info cache due to event network-changed-72671c1a-eabd-4a80-a591-17b26cea0249. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 941.831715] env[62521]: DEBUG oslo_concurrency.lockutils [req-b81858ae-e7d2-4fe9-82a0-f0c30bd3ba3b req-1a51c932-f914-4fef-8525-25ec1fedcc28 service nova] Acquiring lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.831864] env[62521]: DEBUG oslo_concurrency.lockutils [req-b81858ae-e7d2-4fe9-82a0-f0c30bd3ba3b req-1a51c932-f914-4fef-8525-25ec1fedcc28 service nova] Acquired lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.832036] env[62521]: DEBUG nova.network.neutron [req-b81858ae-e7d2-4fe9-82a0-f0c30bd3ba3b req-1a51c932-f914-4fef-8525-25ec1fedcc28 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Refreshing network info cache for port 72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 941.844681] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318817, 'name': ReconfigVM_Task, 'duration_secs': 0.436617} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.844943] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c/02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.845553] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5b57bfcf-b7bd-4b3f-a6bb-7c0855d5242c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.851555] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 941.851555] env[62521]: value = "task-1318819" [ 941.851555] env[62521]: _type = "Task" [ 941.851555] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.856477] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.860006] env[62521]: DEBUG nova.scheduler.client.report [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.863033] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318819, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.930050] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e566ff-d1cb-ecf6-6357-5887e9ad47a3, 'name': SearchDatastore_Task, 'duration_secs': 0.035669} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.933253] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.933566] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] ac420a77-2160-45d6-b313-c5304536c39d/ac420a77-2160-45d6-b313-c5304536c39d.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 941.933847] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8aa680c1-89b2-4125-a2ee-4f08e8ebb034 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.941140] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318818, 'name': CreateVM_Task, 'duration_secs': 0.326069} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.942298] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.942619] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 941.942619] env[62521]: value = "task-1318820" [ 941.942619] env[62521]: _type = "Task" [ 941.942619] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.943236] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.943412] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.943794] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.944097] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28e3a299-1d48-45f3-9eba-04960870efd9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.950922] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 941.950922] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c1b277-baad-91fb-6a9d-63d528163e64" [ 941.950922] env[62521]: _type = "Task" [ 941.950922] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.953868] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318820, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.960978] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c1b277-baad-91fb-6a9d-63d528163e64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.007402] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.007624] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.007890] env[62521]: INFO nova.compute.manager [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Rebooting instance [ 942.025738] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.144336] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d47d6f6-316c-45af-a456-5611d6e0c31d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.164930] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd25107-3a8e-41cd-9738-2e9c069193be {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.172467] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance '17876bcc-3a7b-47b0-afd9-1215e24864ae' progress to 83 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 942.367315] env[62521]: DEBUG oslo_concurrency.lockutils [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.368143] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318819, 'name': Rename_Task, 'duration_secs': 0.144204} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.368832] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.512s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.369235] env[62521]: DEBUG nova.objects.instance [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lazy-loading 'resources' on Instance uuid 3afba9cd-f66e-44c5-b660-b4314f370901 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.372222] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.377226] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd741b4e-5802-48e0-83c6-9d9ee3c273c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.386767] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 942.386767] env[62521]: value = "task-1318821" [ 942.386767] env[62521]: _type = "Task" [ 942.386767] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.400756] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318821, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.401894] env[62521]: INFO nova.scheduler.client.report [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Deleted allocations for instance cc4b29ee-969d-49c6-9fae-71d5f850aae1 [ 942.454357] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318820, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.464042] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.464386] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Processing image 720eb7fe-5500-4cf9-a06e-3471ca9179ed {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.464680] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed/720eb7fe-5500-4cf9-a06e-3471ca9179ed.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.464876] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed/720eb7fe-5500-4cf9-a06e-3471ca9179ed.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.465115] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.465738] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15773655-6ac9-4215-819a-201625325c55 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.473709] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.473896] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 942.474646] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcd12acf-430a-4630-822e-45f7df1ae924 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.482409] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 942.482409] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524a533b-0792-c408-cbec-0d93337a43b8" [ 942.482409] env[62521]: _type = "Task" [ 942.482409] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.490713] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524a533b-0792-c408-cbec-0d93337a43b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.525204] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.525461] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.525717] env[62521]: DEBUG nova.network.neutron [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.631509] env[62521]: DEBUG nova.network.neutron [req-b81858ae-e7d2-4fe9-82a0-f0c30bd3ba3b req-1a51c932-f914-4fef-8525-25ec1fedcc28 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updated VIF entry in instance network info cache for port 72671c1a-eabd-4a80-a591-17b26cea0249. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.631938] env[62521]: DEBUG nova.network.neutron [req-b81858ae-e7d2-4fe9-82a0-f0c30bd3ba3b req-1a51c932-f914-4fef-8525-25ec1fedcc28 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating instance_info_cache with network_info: [{"id": "72671c1a-eabd-4a80-a591-17b26cea0249", "address": "fa:16:3e:3a:1f:96", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72671c1a-ea", "ovs_interfaceid": "72671c1a-eabd-4a80-a591-17b26cea0249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.679321] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.679620] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fb85173-31e9-4bce-84df-7e8d5d2ed613 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.686784] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 942.686784] env[62521]: value = "task-1318822" [ 942.686784] env[62521]: _type = "Task" [ 942.686784] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.701029] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318822, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.898529] env[62521]: DEBUG oslo_vmware.api [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318821, 'name': PowerOnVM_Task, 'duration_secs': 0.46302} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.898860] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 942.899091] env[62521]: INFO nova.compute.manager [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Took 8.31 seconds to spawn the instance on the hypervisor. [ 942.899280] env[62521]: DEBUG nova.compute.manager [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.900636] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c9d99ae-d65c-4471-9127-27d26380d2e6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.912561] env[62521]: DEBUG oslo_concurrency.lockutils [None req-eaa7f791-96fe-4f49-bf65-3488b362f15f tempest-ImagesOneServerNegativeTestJSON-1772686361 tempest-ImagesOneServerNegativeTestJSON-1772686361-project-member] Lock "cc4b29ee-969d-49c6-9fae-71d5f850aae1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.318s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.955123] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318820, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54756} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.955392] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] ac420a77-2160-45d6-b313-c5304536c39d/ac420a77-2160-45d6-b313-c5304536c39d.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 942.955611] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.955860] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bfe639ae-e4fb-4d43-b128-f4c6cd0025f6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.961965] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 942.961965] env[62521]: value = "task-1318823" [ 942.961965] env[62521]: _type = "Task" [ 942.961965] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.971556] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318823, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.995677] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Preparing fetch location {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 942.995967] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Fetch image to [datastore1] OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6/OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6.vmdk {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 942.996192] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Downloading stream optimized image 720eb7fe-5500-4cf9-a06e-3471ca9179ed to [datastore1] OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6/OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6.vmdk on the data store datastore1 as vApp {{(pid=62521) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 942.996408] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Downloading image file data 720eb7fe-5500-4cf9-a06e-3471ca9179ed to the ESX as VM named 'OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6' {{(pid=62521) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 943.079073] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3af3d3-70f7-4fa0-ad66-5ead355e6e31 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.086512] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f70e9f-dace-4cf3-8d56-04712d7ed532 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.092237] env[62521]: DEBUG oslo_vmware.rw_handles [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 943.092237] env[62521]: value = "resgroup-9" [ 943.092237] env[62521]: _type = "ResourcePool" [ 943.092237] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 943.092440] env[62521]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7fdeaebd-f60a-4c91-9b26-13dd7bfeb03e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.134652] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3485789-be58-460f-842d-47958477c5f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.138578] env[62521]: DEBUG oslo_concurrency.lockutils [req-b81858ae-e7d2-4fe9-82a0-f0c30bd3ba3b req-1a51c932-f914-4fef-8525-25ec1fedcc28 service nova] Releasing lock "refresh_cache-73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.138875] env[62521]: DEBUG nova.compute.manager [req-b81858ae-e7d2-4fe9-82a0-f0c30bd3ba3b req-1a51c932-f914-4fef-8525-25ec1fedcc28 service nova] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Received event network-vif-deleted-64c6f406-e4c3-4cae-adb6-5aa4b1c1b9ee {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.139328] env[62521]: DEBUG oslo_vmware.rw_handles [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lease: (returnval){ [ 943.139328] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529dba74-bd80-df62-7c2e-24a3fc879135" [ 943.139328] env[62521]: _type = "HttpNfcLease" [ 943.139328] env[62521]: } obtained for vApp import into resource pool (val){ [ 943.139328] env[62521]: value = "resgroup-9" [ 943.139328] env[62521]: _type = "ResourcePool" [ 943.139328] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 943.139545] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the lease: (returnval){ [ 943.139545] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529dba74-bd80-df62-7c2e-24a3fc879135" [ 943.139545] env[62521]: _type = "HttpNfcLease" [ 943.139545] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 943.145805] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20881ad2-625a-4133-978d-616b3a9a5c76 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.151045] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 943.151045] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529dba74-bd80-df62-7c2e-24a3fc879135" [ 943.151045] env[62521]: _type = "HttpNfcLease" [ 943.151045] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 943.160404] env[62521]: DEBUG nova.compute.provider_tree [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.197326] env[62521]: DEBUG oslo_vmware.api [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318822, 'name': PowerOnVM_Task, 'duration_secs': 0.390549} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.199844] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 943.200072] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-070a8ddb-af32-44e8-a8cc-f84944fa458e tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance '17876bcc-3a7b-47b0-afd9-1215e24864ae' progress to 100 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 943.401153] env[62521]: DEBUG nova.network.neutron [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [{"id": "8824a231-01df-41e5-9bea-9071d8b41cb8", "address": "fa:16:3e:2e:cc:1b", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8824a231-01", "ovs_interfaceid": "8824a231-01df-41e5-9bea-9071d8b41cb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.425159] env[62521]: INFO nova.compute.manager [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Took 15.86 seconds to build instance. [ 943.475678] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318823, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072484} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.476071] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.477199] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5b4cd9-d814-44f6-b918-127d894df8ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.507067] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] ac420a77-2160-45d6-b313-c5304536c39d/ac420a77-2160-45d6-b313-c5304536c39d.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.507836] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb83144f-d431-41a0-89e2-3b392d66791b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.527497] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 943.527497] env[62521]: value = "task-1318825" [ 943.527497] env[62521]: _type = "Task" [ 943.527497] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.535444] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318825, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.648913] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 943.648913] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529dba74-bd80-df62-7c2e-24a3fc879135" [ 943.648913] env[62521]: _type = "HttpNfcLease" [ 943.648913] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 943.663635] env[62521]: DEBUG nova.scheduler.client.report [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.903882] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.906685] env[62521]: DEBUG nova.compute.manager [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.907623] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c571bac-75ba-4154-b013-dc875d9a0a17 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.929601] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8f38b911-3ce0-497a-8ac2-c853ed80a722 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.375s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.042341] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318825, 'name': ReconfigVM_Task, 'duration_secs': 0.412572} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.042341] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Reconfigured VM instance instance-00000053 to attach disk [datastore1] ac420a77-2160-45d6-b313-c5304536c39d/ac420a77-2160-45d6-b313-c5304536c39d.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.042341] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e859653-8bd4-464c-8156-ad06698a68cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.048421] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 944.048421] env[62521]: value = "task-1318826" [ 944.048421] env[62521]: _type = "Task" [ 944.048421] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.061784] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318826, 'name': Rename_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.148974] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 944.148974] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529dba74-bd80-df62-7c2e-24a3fc879135" [ 944.148974] env[62521]: _type = "HttpNfcLease" [ 944.148974] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 944.149394] env[62521]: DEBUG oslo_vmware.rw_handles [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 944.149394] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529dba74-bd80-df62-7c2e-24a3fc879135" [ 944.149394] env[62521]: _type = "HttpNfcLease" [ 944.149394] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 944.150038] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91f36e0-c1ee-40b5-872f-6975105ee7cb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.159615] env[62521]: DEBUG oslo_vmware.rw_handles [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecdd1a-464c-17db-6682-a5dd1036c156/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 944.159615] env[62521]: DEBUG oslo_vmware.rw_handles [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecdd1a-464c-17db-6682-a5dd1036c156/disk-0.vmdk. {{(pid=62521) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 944.218568] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.849s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.229041] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.203s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.229371] env[62521]: DEBUG nova.objects.instance [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lazy-loading 'pci_requests' on Instance uuid bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.240820] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-12798c75-a4a3-4155-89c2-2214b32037ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.270249] env[62521]: INFO nova.scheduler.client.report [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Deleted allocations for instance 3afba9cd-f66e-44c5-b660-b4314f370901 [ 944.560722] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318826, 'name': Rename_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.742642] env[62521]: DEBUG nova.objects.instance [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lazy-loading 'numa_topology' on Instance uuid bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.781126] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bdbc0061-e859-431e-afc6-ffae3a10eb5e tempest-ImagesTestJSON-363662827 tempest-ImagesTestJSON-363662827-project-member] Lock "3afba9cd-f66e-44c5-b660-b4314f370901" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.931s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.934185] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4733c061-2b7a-4b25-8a71-0abf41d4d3d4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.945816] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Doing hard reboot of VM {{(pid=62521) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 944.947062] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-6ceb52d7-67dd-468c-9cac-0c5ed6dc9577 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.955194] env[62521]: DEBUG oslo_vmware.api [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 944.955194] env[62521]: value = "task-1318827" [ 944.955194] env[62521]: _type = "Task" [ 944.955194] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.966246] env[62521]: DEBUG oslo_vmware.api [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318827, 'name': ResetVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.061609] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318826, 'name': Rename_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.244715] env[62521]: INFO nova.compute.claims [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.467534] env[62521]: DEBUG oslo_vmware.api [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318827, 'name': ResetVM_Task, 'duration_secs': 0.098938} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.467807] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Did hard reboot of VM {{(pid=62521) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 945.468101] env[62521]: DEBUG nova.compute.manager [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.469162] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64e744e-71ab-4824-8a96-fc1a9ba80c28 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.533598] env[62521]: DEBUG oslo_vmware.rw_handles [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Completed reading data from the image iterator. {{(pid=62521) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 945.533853] env[62521]: DEBUG oslo_vmware.rw_handles [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecdd1a-464c-17db-6682-a5dd1036c156/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 945.534923] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a134005-1d44-4599-8b1a-086f2762d4ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.543465] env[62521]: DEBUG oslo_vmware.rw_handles [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecdd1a-464c-17db-6682-a5dd1036c156/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 945.543465] env[62521]: DEBUG oslo_vmware.rw_handles [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecdd1a-464c-17db-6682-a5dd1036c156/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 945.543465] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-3dd58da3-793f-4a09-9eda-53b9882594a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.558948] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318826, 'name': Rename_Task, 'duration_secs': 1.165245} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.559326] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.559739] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cda685ff-18dc-4e31-be99-edb14743aaff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.566908] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 945.566908] env[62521]: value = "task-1318828" [ 945.566908] env[62521]: _type = "Task" [ 945.566908] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.576555] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318828, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.641542] env[62521]: DEBUG nova.compute.manager [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Stashing vm_state: active {{(pid=62521) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 945.685126] env[62521]: DEBUG nova.network.neutron [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Port 3b0934da-f42d-4e91-ae7c-a4f254d6b86c binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 945.685126] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.685299] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.685501] env[62521]: DEBUG nova.network.neutron [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 945.785149] env[62521]: DEBUG oslo_vmware.rw_handles [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecdd1a-464c-17db-6682-a5dd1036c156/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 945.785459] env[62521]: INFO nova.virt.vmwareapi.images [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Downloaded image file data 720eb7fe-5500-4cf9-a06e-3471ca9179ed [ 945.786397] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65455602-121a-44e2-8292-1212cf19fdde {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.806659] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1753c044-814e-4c83-9928-02076dea1216 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.839793] env[62521]: INFO nova.virt.vmwareapi.images [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] The imported VM was unregistered [ 945.842964] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Caching image {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 945.843425] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating directory with path [datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.843805] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b3ac2d4-b921-42e8-9e8a-ac927c004c01 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.859407] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Created directory with path [datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.859669] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6/OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6.vmdk to [datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed/720eb7fe-5500-4cf9-a06e-3471ca9179ed.vmdk. {{(pid=62521) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 945.859984] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-9458a76a-52ee-45df-a7c8-d39f8271b534 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.871793] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 945.871793] env[62521]: value = "task-1318830" [ 945.871793] env[62521]: _type = "Task" [ 945.871793] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.880967] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318830, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.986951] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c549f254-4b86-4193-9443-c9ddce1c2950 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.979s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.079841] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318828, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.171020] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.386394] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318830, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.476916] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdda4c79-7b2a-4032-a584-208feb120987 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.484984] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6d6e8d-9e5e-461c-9ec1-166dbba0be0f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.517824] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2118a2e2-58f5-4856-b2c5-99b892f2b600 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.526577] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79b8ec2-1d56-4e29-a9e9-0317136c0ecb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.542288] env[62521]: DEBUG nova.compute.provider_tree [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.579762] env[62521]: DEBUG oslo_vmware.api [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318828, 'name': PowerOnVM_Task, 'duration_secs': 0.578604} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.580241] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 946.580665] env[62521]: INFO nova.compute.manager [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Took 9.66 seconds to spawn the instance on the hypervisor. [ 946.580938] env[62521]: DEBUG nova.compute.manager [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.582291] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d223ec80-2bb3-4138-906a-eaae5fa344a6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.676680] env[62521]: DEBUG nova.network.neutron [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance_info_cache with network_info: [{"id": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "address": "fa:16:3e:1a:80:af", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0934da-f4", "ovs_interfaceid": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.884429] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318830, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.045801] env[62521]: DEBUG nova.scheduler.client.report [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.115193] env[62521]: INFO nova.compute.manager [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Took 16.46 seconds to build instance. [ 947.181431] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.387283] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318830, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.469208] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "563ad3b2-e53e-4682-9cde-bd6f709718de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.469488] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "563ad3b2-e53e-4682-9cde-bd6f709718de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.469740] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "563ad3b2-e53e-4682-9cde-bd6f709718de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.469960] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "563ad3b2-e53e-4682-9cde-bd6f709718de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.470197] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "563ad3b2-e53e-4682-9cde-bd6f709718de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.478023] env[62521]: INFO nova.compute.manager [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Terminating instance [ 947.478023] env[62521]: DEBUG nova.compute.manager [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.478023] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 947.478237] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3978b392-9a42-4e9e-9422-c0b0ca082956 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.488174] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 947.488174] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfdfd19d-0a89-4baa-9d0f-994f1e460d99 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.495759] env[62521]: DEBUG oslo_vmware.api [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 947.495759] env[62521]: value = "task-1318831" [ 947.495759] env[62521]: _type = "Task" [ 947.495759] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.506335] env[62521]: DEBUG oslo_vmware.api [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.553037] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.324s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.556555] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.385s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.607812] env[62521]: INFO nova.network.neutron [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updating port 903b74d0-11d2-404d-81f1-b6bb18f713af with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 947.616838] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ff73f44a-66a4-46f7-aa90-8700a6c47284 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "ac420a77-2160-45d6-b313-c5304536c39d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.982s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.685351] env[62521]: DEBUG nova.compute.manager [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62521) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 947.685560] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.885902] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318830, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.005471] env[62521]: DEBUG oslo_vmware.api [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318831, 'name': PowerOffVM_Task, 'duration_secs': 0.315857} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.005826] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 948.006048] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 948.006944] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0a4c6a3-44b3-467f-aaf8-41a6a4dcd8af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.063267] env[62521]: INFO nova.compute.claims [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 948.096331] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 948.096614] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 948.096807] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleting the datastore file [datastore1] 563ad3b2-e53e-4682-9cde-bd6f709718de {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.097101] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b10b38d-61db-48f7-8dee-8f64b1fb6d93 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.108272] env[62521]: DEBUG oslo_vmware.api [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 948.108272] env[62521]: value = "task-1318833" [ 948.108272] env[62521]: _type = "Task" [ 948.108272] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.123218] env[62521]: DEBUG oslo_vmware.api [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.267389] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "ac420a77-2160-45d6-b313-c5304536c39d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.267389] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "ac420a77-2160-45d6-b313-c5304536c39d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.267389] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "ac420a77-2160-45d6-b313-c5304536c39d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.267570] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "ac420a77-2160-45d6-b313-c5304536c39d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.267863] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "ac420a77-2160-45d6-b313-c5304536c39d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.270824] env[62521]: INFO nova.compute.manager [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Terminating instance [ 948.274030] env[62521]: DEBUG nova.compute.manager [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 948.274125] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 948.275682] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba155d58-583f-4a38-9cb6-680e5f12fa9d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.283628] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.283915] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f19ff1e-1143-4c4d-a5ae-6fc6ceafc743 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.293597] env[62521]: DEBUG oslo_vmware.api [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 948.293597] env[62521]: value = "task-1318834" [ 948.293597] env[62521]: _type = "Task" [ 948.293597] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.301859] env[62521]: DEBUG oslo_vmware.api [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318834, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.327626] env[62521]: DEBUG nova.compute.manager [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.328715] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddbbaf3-3edd-4c42-a523-68685279a7af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.392570] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318830, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.576868] env[62521]: INFO nova.compute.resource_tracker [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating resource usage from migration 371d8a8a-aab8-45be-aa37-32f199c69c75 [ 948.619223] env[62521]: DEBUG oslo_vmware.api [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.803722] env[62521]: DEBUG oslo_vmware.api [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318834, 'name': PowerOffVM_Task, 'duration_secs': 0.452892} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.804074] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 948.804192] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 948.804472] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fee4ea2-9364-4bc7-9e17-e2127f97156c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.811107] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35fe2c1-7d57-4a8a-8125-0db057553414 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.817213] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe97a1bf-019c-4e64-950d-c47923a21dc3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.856678] env[62521]: INFO nova.compute.manager [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] instance snapshotting [ 948.857333] env[62521]: DEBUG nova.objects.instance [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'flavor' on Instance uuid 11e61e72-a311-4c43-bf53-df82ad9d70f7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.859747] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb307281-e298-41dc-8e61-4ea45d3c37ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.869313] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 948.869543] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 948.869706] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleting the datastore file [datastore1] ac420a77-2160-45d6-b313-c5304536c39d {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.871900] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18e3dfaa-cff4-4c3a-8ee7-9f4d1a92f95b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.877098] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ec6922-7d14-424b-bf71-fc25eec31e5b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.885223] env[62521]: DEBUG oslo_vmware.api [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 948.885223] env[62521]: value = "task-1318836" [ 948.885223] env[62521]: _type = "Task" [ 948.885223] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.895471] env[62521]: DEBUG nova.compute.provider_tree [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.900965] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318830, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.905780] env[62521]: DEBUG oslo_vmware.api [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318836, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.121378] env[62521]: DEBUG oslo_vmware.api [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.369022] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1105b56-dc43-40fc-8538-df974bf10fde {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.410415] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f9f63a-9501-43fc-baf7-364bc6bc1de1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.413712] env[62521]: DEBUG nova.scheduler.client.report [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.422149] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318830, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.201634} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.427401] env[62521]: INFO nova.virt.vmwareapi.ds_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6/OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6.vmdk to [datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed/720eb7fe-5500-4cf9-a06e-3471ca9179ed.vmdk. [ 949.427863] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Cleaning up location [datastore1] OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6 {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 949.428202] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_7fc9a21d-5374-4c0f-b488-4213f93051b6 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.433854] env[62521]: DEBUG oslo_vmware.api [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318836, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.433854] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f6a9e35-0fad-43be-8627-cb913f31dc50 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.440019] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 949.440019] env[62521]: value = "task-1318837" [ 949.440019] env[62521]: _type = "Task" [ 949.440019] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.456402] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318837, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.619899] env[62521]: DEBUG oslo_vmware.api [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318833, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.106116} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.620182] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 949.620372] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 949.620582] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 949.620778] env[62521]: INFO nova.compute.manager [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Took 2.14 seconds to destroy the instance on the hypervisor. [ 949.621063] env[62521]: DEBUG oslo.service.loopingcall [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.621289] env[62521]: DEBUG nova.compute.manager [-] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 949.621389] env[62521]: DEBUG nova.network.neutron [-] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 949.635185] env[62521]: DEBUG nova.compute.manager [req-f09c965f-7fd0-496e-8c4d-8f4fb57e4c0b req-baee6008-9133-494a-863a-daa85c44bfed service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Received event network-vif-plugged-903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.635185] env[62521]: DEBUG oslo_concurrency.lockutils [req-f09c965f-7fd0-496e-8c4d-8f4fb57e4c0b req-baee6008-9133-494a-863a-daa85c44bfed service nova] Acquiring lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.635185] env[62521]: DEBUG oslo_concurrency.lockutils [req-f09c965f-7fd0-496e-8c4d-8f4fb57e4c0b req-baee6008-9133-494a-863a-daa85c44bfed service nova] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.635185] env[62521]: DEBUG oslo_concurrency.lockutils [req-f09c965f-7fd0-496e-8c4d-8f4fb57e4c0b req-baee6008-9133-494a-863a-daa85c44bfed service nova] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.635659] env[62521]: DEBUG nova.compute.manager [req-f09c965f-7fd0-496e-8c4d-8f4fb57e4c0b req-baee6008-9133-494a-863a-daa85c44bfed service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] No waiting events found dispatching network-vif-plugged-903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 949.636035] env[62521]: WARNING nova.compute.manager [req-f09c965f-7fd0-496e-8c4d-8f4fb57e4c0b req-baee6008-9133-494a-863a-daa85c44bfed service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Received unexpected event network-vif-plugged-903b74d0-11d2-404d-81f1-b6bb18f713af for instance with vm_state shelved_offloaded and task_state spawning. [ 949.698685] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.698890] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.699079] env[62521]: DEBUG nova.network.neutron [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 949.801295] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.801530] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.910544] env[62521]: DEBUG oslo_vmware.api [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318836, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.910012} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.911568] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 949.911568] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 949.911568] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 949.911568] env[62521]: INFO nova.compute.manager [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Took 1.64 seconds to destroy the instance on the hypervisor. [ 949.911770] env[62521]: DEBUG oslo.service.loopingcall [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.911811] env[62521]: DEBUG nova.compute.manager [-] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 949.912493] env[62521]: DEBUG nova.network.neutron [-] [instance: ac420a77-2160-45d6-b313-c5304536c39d] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 949.919521] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.363s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.919722] env[62521]: INFO nova.compute.manager [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Migrating [ 949.927605] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 2.242s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.935626] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 949.941870] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-85dac1e0-227c-4f0e-87b8-59ca5ef61ed6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.958979] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 949.958979] env[62521]: value = "task-1318838" [ 949.958979] env[62521]: _type = "Task" [ 949.958979] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.960238] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318837, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170738} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.960238] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 949.961242] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed/720eb7fe-5500-4cf9-a06e-3471ca9179ed.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.961242] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed/720eb7fe-5500-4cf9-a06e-3471ca9179ed.vmdk to [datastore1] 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad/73bc1cf1-7ee1-4a9b-8214-2dc752b700ad.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.967022] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cac79b31-846b-43e2-92d4-027898e8fe44 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.974634] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318838, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.976337] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 949.976337] env[62521]: value = "task-1318839" [ 949.976337] env[62521]: _type = "Task" [ 949.976337] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.993018] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.251242] env[62521]: DEBUG nova.compute.manager [req-f59552d7-088d-4c29-90c1-ff11cb714576 req-5e996960-5410-4769-9429-f84ec32249a9 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Received event network-vif-deleted-2c0a3f29-63a4-44ff-bbf5-40908d17dc42 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.251242] env[62521]: INFO nova.compute.manager [req-f59552d7-088d-4c29-90c1-ff11cb714576 req-5e996960-5410-4769-9429-f84ec32249a9 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Neutron deleted interface 2c0a3f29-63a4-44ff-bbf5-40908d17dc42; detaching it from the instance and deleting it from the info cache [ 950.251242] env[62521]: DEBUG nova.network.neutron [req-f59552d7-088d-4c29-90c1-ff11cb714576 req-5e996960-5410-4769-9429-f84ec32249a9 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.303571] env[62521]: DEBUG nova.compute.manager [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 950.435440] env[62521]: DEBUG nova.objects.instance [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lazy-loading 'migration_context' on Instance uuid 17876bcc-3a7b-47b0-afd9-1215e24864ae {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.443149] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.443617] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.444072] env[62521]: DEBUG nova.network.neutron [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.471164] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318838, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.486096] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318839, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.544888] env[62521]: DEBUG nova.network.neutron [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updating instance_info_cache with network_info: [{"id": "903b74d0-11d2-404d-81f1-b6bb18f713af", "address": "fa:16:3e:6a:82:e1", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap903b74d0-11", "ovs_interfaceid": "903b74d0-11d2-404d-81f1-b6bb18f713af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.694240] env[62521]: DEBUG nova.network.neutron [-] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.755880] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-84e4ada3-18b2-41d2-b55c-c315ef85943b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.769799] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f555f5a-d90b-4bde-946a-b6a64c4f0264 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.799715] env[62521]: DEBUG nova.compute.manager [req-f59552d7-088d-4c29-90c1-ff11cb714576 req-5e996960-5410-4769-9429-f84ec32249a9 service nova] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Detach interface failed, port_id=2c0a3f29-63a4-44ff-bbf5-40908d17dc42, reason: Instance 563ad3b2-e53e-4682-9cde-bd6f709718de could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 950.832759] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.842345] env[62521]: DEBUG nova.network.neutron [-] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.974261] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318838, 'name': CreateSnapshot_Task, 'duration_secs': 0.737826} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.974582] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 950.975398] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e20c243-c458-4199-a461-2d04c4274fbb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.992177] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318839, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.048980] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.076497] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='39ce3692b2ead274340362e1419956dc',container_format='bare',created_at=2024-10-20T00:36:40Z,direct_url=,disk_format='vmdk',id=2919854a-4a07-48e6-87af-29f13a5c2313,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1758271622-shelved',owner='a592d0f3c5834858bda6ca89a33662a9',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-20T00:36:56Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.076825] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.077017] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.077147] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.077310] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.077394] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.077658] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.077775] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.078065] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.078134] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.078312] env[62521]: DEBUG nova.virt.hardware [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.079263] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188ec4bf-48ec-4d86-8d00-b1625b0fbe33 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.092347] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18d67ef-c1a1-4497-b8c8-dc6c3a1bbb53 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.108443] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:82:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd646f9d5-d2ad-4c22-bea5-85a965334de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '903b74d0-11d2-404d-81f1-b6bb18f713af', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.116259] env[62521]: DEBUG oslo.service.loopingcall [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.116595] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.116970] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75a80fee-c712-4564-8be5-c488ea2f43e9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.143174] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.143174] env[62521]: value = "task-1318840" [ 951.143174] env[62521]: _type = "Task" [ 951.143174] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.151027] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318840, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.196329] env[62521]: INFO nova.compute.manager [-] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Took 1.57 seconds to deallocate network for instance. [ 951.255198] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eaac5d0-9f6e-4d54-a888-bdb407f78db5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.263942] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971f7330-7ee8-4711-8978-b4372d8acd1c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.295932] env[62521]: DEBUG nova.network.neutron [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance_info_cache with network_info: [{"id": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "address": "fa:16:3e:8c:86:c8", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34cee820-50", "ovs_interfaceid": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.298367] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca53329a-97dd-4fb9-8e7a-629bc71145fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.306976] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-838504e3-fa60-4669-9330-f26c1caaa2c8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.324717] env[62521]: DEBUG nova.compute.provider_tree [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.344556] env[62521]: INFO nova.compute.manager [-] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Took 1.43 seconds to deallocate network for instance. [ 951.489663] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318839, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.506656] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 951.507232] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-602603f4-db78-42d1-b07e-56b2c18b5a16 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.516491] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 951.516491] env[62521]: value = "task-1318841" [ 951.516491] env[62521]: _type = "Task" [ 951.516491] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.526756] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318841, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.653750] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318840, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.703687] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.727201] env[62521]: DEBUG nova.compute.manager [req-6e4b2a11-3492-4ab8-97a1-c5d4e20793e2 req-56890ecf-d00d-406a-aa70-3838cd45b20c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Received event network-changed-903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.727451] env[62521]: DEBUG nova.compute.manager [req-6e4b2a11-3492-4ab8-97a1-c5d4e20793e2 req-56890ecf-d00d-406a-aa70-3838cd45b20c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Refreshing instance network info cache due to event network-changed-903b74d0-11d2-404d-81f1-b6bb18f713af. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 951.727666] env[62521]: DEBUG oslo_concurrency.lockutils [req-6e4b2a11-3492-4ab8-97a1-c5d4e20793e2 req-56890ecf-d00d-406a-aa70-3838cd45b20c service nova] Acquiring lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.727871] env[62521]: DEBUG oslo_concurrency.lockutils [req-6e4b2a11-3492-4ab8-97a1-c5d4e20793e2 req-56890ecf-d00d-406a-aa70-3838cd45b20c service nova] Acquired lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.727980] env[62521]: DEBUG nova.network.neutron [req-6e4b2a11-3492-4ab8-97a1-c5d4e20793e2 req-56890ecf-d00d-406a-aa70-3838cd45b20c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Refreshing network info cache for port 903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.802927] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.828236] env[62521]: DEBUG nova.scheduler.client.report [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.852182] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.988857] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318839, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.028897] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318841, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.156389] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318840, 'name': CreateVM_Task, 'duration_secs': 0.902887} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.156629] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 952.157403] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.157605] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.158038] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.158395] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ae5baf7-a460-473a-ac20-886af31f735c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.164107] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 952.164107] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52384d05-d5e9-1a1f-2650-87aa83721cad" [ 952.164107] env[62521]: _type = "Task" [ 952.164107] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.172876] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52384d05-d5e9-1a1f-2650-87aa83721cad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.494712] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318839, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.498228} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.495383] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/720eb7fe-5500-4cf9-a06e-3471ca9179ed/720eb7fe-5500-4cf9-a06e-3471ca9179ed.vmdk to [datastore1] 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad/73bc1cf1-7ee1-4a9b-8214-2dc752b700ad.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.496368] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c182fe-c0ee-4ee4-9c95-a5c4aed09dc9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.520198] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad/73bc1cf1-7ee1-4a9b-8214-2dc752b700ad.vmdk or device None with type streamOptimized {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.524023] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a21a8f34-0b62-4736-857d-a197a323c760 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.547107] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318841, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.548362] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 952.548362] env[62521]: value = "task-1318842" [ 952.548362] env[62521]: _type = "Task" [ 952.548362] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.557082] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318842, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.606274] env[62521]: DEBUG nova.network.neutron [req-6e4b2a11-3492-4ab8-97a1-c5d4e20793e2 req-56890ecf-d00d-406a-aa70-3838cd45b20c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updated VIF entry in instance network info cache for port 903b74d0-11d2-404d-81f1-b6bb18f713af. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 952.606274] env[62521]: DEBUG nova.network.neutron [req-6e4b2a11-3492-4ab8-97a1-c5d4e20793e2 req-56890ecf-d00d-406a-aa70-3838cd45b20c service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updating instance_info_cache with network_info: [{"id": "903b74d0-11d2-404d-81f1-b6bb18f713af", "address": "fa:16:3e:6a:82:e1", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap903b74d0-11", "ovs_interfaceid": "903b74d0-11d2-404d-81f1-b6bb18f713af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.679959] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.680382] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Processing image 2919854a-4a07-48e6-87af-29f13a5c2313 {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.680942] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313/2919854a-4a07-48e6-87af-29f13a5c2313.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.681189] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313/2919854a-4a07-48e6-87af-29f13a5c2313.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.681466] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 952.681951] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a7f4b8f-ea7e-4c7d-bf1a-7fdf734b1750 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.693782] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 952.693979] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 952.694783] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-173846f5-bf5b-42e6-9440-18ab3eaa747d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.700017] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 952.700017] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b34992-15ab-2441-2c07-3d468a58e1bf" [ 952.700017] env[62521]: _type = "Task" [ 952.700017] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.710565] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b34992-15ab-2441-2c07-3d468a58e1bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.841844] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.914s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.854817] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.022s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.857086] env[62521]: INFO nova.compute.claims [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 953.033149] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318841, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.059046] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318842, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.109947] env[62521]: DEBUG oslo_concurrency.lockutils [req-6e4b2a11-3492-4ab8-97a1-c5d4e20793e2 req-56890ecf-d00d-406a-aa70-3838cd45b20c service nova] Releasing lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.110214] env[62521]: DEBUG nova.compute.manager [req-6e4b2a11-3492-4ab8-97a1-c5d4e20793e2 req-56890ecf-d00d-406a-aa70-3838cd45b20c service nova] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Received event network-vif-deleted-00b03967-a3d2-4262-9736-6cef5c454664 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.210803] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Preparing fetch location {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 953.211116] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Fetch image to [datastore1] OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601/OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601.vmdk {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 953.211621] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Downloading stream optimized image 2919854a-4a07-48e6-87af-29f13a5c2313 to [datastore1] OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601/OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601.vmdk on the data store datastore1 as vApp {{(pid=62521) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 953.211621] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Downloading image file data 2919854a-4a07-48e6-87af-29f13a5c2313 to the ESX as VM named 'OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601' {{(pid=62521) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 953.297154] env[62521]: DEBUG oslo_vmware.rw_handles [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 953.297154] env[62521]: value = "resgroup-9" [ 953.297154] env[62521]: _type = "ResourcePool" [ 953.297154] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 953.298358] env[62521]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-312ea851-e6ae-4358-946c-bf893e50c9ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.322717] env[62521]: DEBUG oslo_vmware.rw_handles [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lease: (returnval){ [ 953.322717] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52626a41-5a13-6c42-b279-798b67d18a13" [ 953.322717] env[62521]: _type = "HttpNfcLease" [ 953.322717] env[62521]: } obtained for vApp import into resource pool (val){ [ 953.322717] env[62521]: value = "resgroup-9" [ 953.322717] env[62521]: _type = "ResourcePool" [ 953.322717] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 953.322983] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the lease: (returnval){ [ 953.322983] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52626a41-5a13-6c42-b279-798b67d18a13" [ 953.322983] env[62521]: _type = "HttpNfcLease" [ 953.322983] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 953.324382] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e31da7-cf5d-4676-bd9a-4ce310dfc790 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.334708] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 953.334708] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52626a41-5a13-6c42-b279-798b67d18a13" [ 953.334708] env[62521]: _type = "HttpNfcLease" [ 953.334708] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 953.348733] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance '02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c' progress to 0 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 953.533766] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318841, 'name': CloneVM_Task, 'duration_secs': 1.921365} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.534089] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Created linked-clone VM from snapshot [ 953.534677] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b858566c-85d2-4d95-93e6-7d02b4d71073 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.542705] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Uploading image 07530e24-0ad3-44e9-9fb4-ee46d23c744c {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 953.557860] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318842, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.575125] env[62521]: DEBUG oslo_vmware.rw_handles [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 953.575125] env[62521]: value = "vm-282145" [ 953.575125] env[62521]: _type = "VirtualMachine" [ 953.575125] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 953.576557] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-21e93f0f-bcb8-42e4-afb6-f5336d68b1d9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.581998] env[62521]: DEBUG oslo_vmware.rw_handles [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lease: (returnval){ [ 953.581998] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52acee79-bead-5e01-90e5-588a437abdb3" [ 953.581998] env[62521]: _type = "HttpNfcLease" [ 953.581998] env[62521]: } obtained for exporting VM: (result){ [ 953.581998] env[62521]: value = "vm-282145" [ 953.581998] env[62521]: _type = "VirtualMachine" [ 953.581998] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 953.582338] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the lease: (returnval){ [ 953.582338] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52acee79-bead-5e01-90e5-588a437abdb3" [ 953.582338] env[62521]: _type = "HttpNfcLease" [ 953.582338] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 953.588952] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 953.588952] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52acee79-bead-5e01-90e5-588a437abdb3" [ 953.588952] env[62521]: _type = "HttpNfcLease" [ 953.588952] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 953.832430] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 953.832430] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52626a41-5a13-6c42-b279-798b67d18a13" [ 953.832430] env[62521]: _type = "HttpNfcLease" [ 953.832430] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 953.856156] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.856156] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-684132e1-03bd-4a5e-ad8f-a84b5a039f10 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.862639] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 953.862639] env[62521]: value = "task-1318845" [ 953.862639] env[62521]: _type = "Task" [ 953.862639] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.879334] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318845, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.063438] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318842, 'name': ReconfigVM_Task, 'duration_secs': 1.302018} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.064431] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad/73bc1cf1-7ee1-4a9b-8214-2dc752b700ad.vmdk or device None with type streamOptimized {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.066468] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'boot_index': 0, 'encryption_secret_uuid': None, 'encryption_format': None, 'device_type': 'disk', 'guest_format': None, 'encryption_options': None, 'disk_bus': None, 'size': 0, 'encrypted': False, 'device_name': '/dev/sda', 'image_id': '58bd9a24-a0a4-4846-b299-475fa0f1d05d'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'delete_on_termination': False, 'device_type': None, 'mount_device': '/dev/sdb', 'guest_format': None, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282136', 'volume_id': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'name': 'volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '73bc1cf1-7ee1-4a9b-8214-2dc752b700ad', 'attached_at': '', 'detached_at': '', 'volume_id': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'serial': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c'}, 'attachment_id': 'ba14b520-08f7-464f-8613-1c2608bdfe0f', 'volume_type': None}], 'swap': None} {{(pid=62521) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 954.067671] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 954.067671] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282136', 'volume_id': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'name': 'volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '73bc1cf1-7ee1-4a9b-8214-2dc752b700ad', 'attached_at': '', 'detached_at': '', 'volume_id': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'serial': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 954.070165] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d041c1e2-6365-4b64-b06c-38a27601e4c1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.102261] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588c0270-9dba-4bce-ad91-26804a617ad0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.108987] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.108987] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52acee79-bead-5e01-90e5-588a437abdb3" [ 954.108987] env[62521]: _type = "HttpNfcLease" [ 954.108987] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 954.124663] env[62521]: DEBUG oslo_vmware.rw_handles [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 954.124663] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52acee79-bead-5e01-90e5-588a437abdb3" [ 954.124663] env[62521]: _type = "HttpNfcLease" [ 954.124663] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 954.132599] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c/volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.136284] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b04f5c8-2f45-4511-8bf6-86bb68361c73 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.138794] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0629a9c-9fca-4181-ae20-b913d1b6a464 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.157856] env[62521]: DEBUG oslo_vmware.rw_handles [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52399f2a-09cc-82b5-feec-8e2f0ca9c197/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 954.158139] env[62521]: DEBUG oslo_vmware.rw_handles [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52399f2a-09cc-82b5-feec-8e2f0ca9c197/disk-0.vmdk for reading. {{(pid=62521) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 954.162461] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 954.162461] env[62521]: value = "task-1318846" [ 954.162461] env[62521]: _type = "Task" [ 954.162461] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.164888] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38558a86-d516-4fc2-9af5-6a8a1ea2ff6f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.233586] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0437ab9-e361-4b1c-bc3e-e4824a3cd463 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.236829] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.267848] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d9fd55-39f8-4f68-8dbf-8f9516ee7657 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.270705] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2b0cd37f-bfee-423a-8062-3f806ae53c3b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.278418] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee351a73-ddfc-4368-8f23-44f38e261c4a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.292999] env[62521]: DEBUG nova.compute.provider_tree [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.332304] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.332304] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52626a41-5a13-6c42-b279-798b67d18a13" [ 954.332304] env[62521]: _type = "HttpNfcLease" [ 954.332304] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 954.332615] env[62521]: DEBUG oslo_vmware.rw_handles [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 954.332615] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52626a41-5a13-6c42-b279-798b67d18a13" [ 954.332615] env[62521]: _type = "HttpNfcLease" [ 954.332615] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 954.333579] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7955c0ae-17ad-4d12-b1e4-21829563d4dd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.340327] env[62521]: DEBUG oslo_vmware.rw_handles [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524fbdb2-b838-6f38-46d2-14dd6cb40f51/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 954.340512] env[62521]: DEBUG oslo_vmware.rw_handles [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524fbdb2-b838-6f38-46d2-14dd6cb40f51/disk-0.vmdk. {{(pid=62521) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 954.406290] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3f88c5dd-2114-4094-bf08-e59b02ed5521 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.412876] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318845, 'name': PowerOffVM_Task, 'duration_secs': 0.295358} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.413387] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.413651] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance '02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c' progress to 17 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 954.421043] env[62521]: INFO nova.compute.manager [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Swapping old allocation on dict_keys(['3d21544b-bfc8-42d0-86ca-d323b5e2628f']) held by migration a367e195-f862-4960-ab2d-b0ece3adece2 for instance [ 954.470240] env[62521]: DEBUG nova.scheduler.client.report [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Overwriting current allocation {'allocations': {'3d21544b-bfc8-42d0-86ca-d323b5e2628f': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 108}}, 'project_id': 'ea161b1af2bb4b9f8cc39b16582f7013', 'user_id': '937840982b914e68849319714a16b1a8', 'consumer_generation': 1} on consumer 17876bcc-3a7b-47b0-afd9-1215e24864ae {{(pid=62521) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 954.554718] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.555083] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquired lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.555147] env[62521]: DEBUG nova.network.neutron [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.679738] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318846, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.800120] env[62521]: DEBUG nova.scheduler.client.report [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.924708] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 954.925183] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 954.925265] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.925460] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 954.925768] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.926120] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 954.926559] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 954.926654] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 954.926949] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 954.927218] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 954.927485] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.940748] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3cad74a6-89d4-448f-bab8-dd21304948e8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.965029] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 954.965029] env[62521]: value = "task-1318847" [ 954.965029] env[62521]: _type = "Task" [ 954.965029] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.973390] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318847, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.176580] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318846, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.322961] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.322961] env[62521]: DEBUG nova.compute.manager [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 955.331053] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.627s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.331203] env[62521]: DEBUG nova.objects.instance [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lazy-loading 'resources' on Instance uuid 563ad3b2-e53e-4682-9cde-bd6f709718de {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.441099] env[62521]: DEBUG nova.network.neutron [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance_info_cache with network_info: [{"id": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "address": "fa:16:3e:1a:80:af", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b0934da-f4", "ovs_interfaceid": "3b0934da-f42d-4e91-ae7c-a4f254d6b86c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.485277] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "b8158051-98b6-47d8-84c5-d1d56515afe8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.485621] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "b8158051-98b6-47d8-84c5-d1d56515afe8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.497996] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318847, 'name': ReconfigVM_Task, 'duration_secs': 0.241606} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.499063] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance '02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c' progress to 33 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 955.680608] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318846, 'name': ReconfigVM_Task, 'duration_secs': 1.374571} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.682543] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Reconfigured VM instance instance-00000042 to attach disk [datastore2] volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c/volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.687633] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b053abd2-2203-4c95-8222-9ecfd2d2e5dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.704098] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 955.704098] env[62521]: value = "task-1318848" [ 955.704098] env[62521]: _type = "Task" [ 955.704098] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.713777] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318848, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.784544] env[62521]: DEBUG oslo_vmware.rw_handles [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Completed reading data from the image iterator. {{(pid=62521) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 955.784869] env[62521]: DEBUG oslo_vmware.rw_handles [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524fbdb2-b838-6f38-46d2-14dd6cb40f51/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 955.786068] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fb8588-6ffd-4346-ba12-d665ef1b722c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.793974] env[62521]: DEBUG oslo_vmware.rw_handles [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524fbdb2-b838-6f38-46d2-14dd6cb40f51/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 955.793974] env[62521]: DEBUG oslo_vmware.rw_handles [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524fbdb2-b838-6f38-46d2-14dd6cb40f51/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 955.794669] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-cf7c9876-7ff1-4826-8d93-da0de4a6b209 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.836331] env[62521]: DEBUG nova.compute.utils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 955.841616] env[62521]: DEBUG nova.compute.manager [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 955.841823] env[62521]: DEBUG nova.network.neutron [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 955.899506] env[62521]: DEBUG nova.policy [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3566601952144168b5d400c41026c81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0dfb5f5bd4646ec95b7c6a2d0434175', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 955.945431] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Releasing lock "refresh_cache-17876bcc-3a7b-47b0-afd9-1215e24864ae" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.945970] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.946661] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba58e864-ac3c-4faf-b07b-087a6da87ebc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.954734] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 955.954734] env[62521]: value = "task-1318849" [ 955.954734] env[62521]: _type = "Task" [ 955.954734] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.967132] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318849, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.999266] env[62521]: DEBUG nova.compute.manager [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 956.007459] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.007671] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.007910] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.008142] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.008306] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.008583] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.008959] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.009195] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.009394] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.009608] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.009903] env[62521]: DEBUG nova.virt.hardware [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.015435] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Reconfiguring VM instance instance-00000052 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 956.016748] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb93aab7-66a9-41b8-8af9-124cd83c1a6a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.040858] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 956.040858] env[62521]: value = "task-1318850" [ 956.040858] env[62521]: _type = "Task" [ 956.040858] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.050850] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318850, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.095976] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e80e2c-737a-4bd0-83fd-a9d824bdb54d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.106460] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff4f0de-3058-4a0d-bb78-7e27c96022ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.117771] env[62521]: DEBUG oslo_vmware.rw_handles [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524fbdb2-b838-6f38-46d2-14dd6cb40f51/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 956.118590] env[62521]: INFO nova.virt.vmwareapi.images [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Downloaded image file data 2919854a-4a07-48e6-87af-29f13a5c2313 [ 956.120757] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4d8a1b-792e-497a-8537-7b77a80c8389 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.157102] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f10950-ab9e-49a9-89d7-a77eb5f3c418 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.173526] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03cc0d54-e027-4ec8-a1e4-44d34ab3e63f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.178240] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc52057-5a78-4105-88a0-7ba1cac8797b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.192791] env[62521]: DEBUG nova.compute.provider_tree [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.202126] env[62521]: INFO nova.virt.vmwareapi.images [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] The imported VM was unregistered [ 956.204280] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Caching image {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 956.204553] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Creating directory with path [datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313 {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.205581] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d15856dd-1d86-448c-bbfe-f6204e311af1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.221313] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318848, 'name': ReconfigVM_Task, 'duration_secs': 0.180809} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.221634] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282136', 'volume_id': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'name': 'volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '73bc1cf1-7ee1-4a9b-8214-2dc752b700ad', 'attached_at': '', 'detached_at': '', 'volume_id': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'serial': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 956.222306] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-598ecdd8-cc38-43d3-844e-7654b5b54a78 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.229039] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 956.229039] env[62521]: value = "task-1318852" [ 956.229039] env[62521]: _type = "Task" [ 956.229039] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.235373] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Created directory with path [datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313 {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.235654] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601/OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601.vmdk to [datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313/2919854a-4a07-48e6-87af-29f13a5c2313.vmdk. {{(pid=62521) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 956.239655] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-bd72bc86-8a2b-4d89-b4f7-05b8dcc2b5dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.241911] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318852, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.250363] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 956.250363] env[62521]: value = "task-1318853" [ 956.250363] env[62521]: _type = "Task" [ 956.250363] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.261196] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318853, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.263640] env[62521]: DEBUG nova.network.neutron [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Successfully created port: fb88c8dc-e610-4e05-ba90-1a6dcf6ae085 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 956.344692] env[62521]: DEBUG nova.compute.manager [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 956.466255] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318849, 'name': PowerOffVM_Task, 'duration_secs': 0.308019} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.466255] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:36:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='e986ebeb-2013-4e50-9994-d8d533ed7ed2',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-778228551',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.469931] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.473330] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73379f31-cfd4-407a-ae26-2c7158aadc2e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.489305] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 956.489305] env[62521]: value = "task-1318854" [ 956.489305] env[62521]: _type = "Task" [ 956.489305] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.497605] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318854, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.535321] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.551682] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318850, 'name': ReconfigVM_Task, 'duration_secs': 0.267905} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.552063] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Reconfigured VM instance instance-00000052 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 956.552925] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c28bcf-e74a-4469-8774-287eec60c9fb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.577184] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c/02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.577459] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ecf772e-1d86-4020-85be-857d6da7c56f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.595149] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 956.595149] env[62521]: value = "task-1318855" [ 956.595149] env[62521]: _type = "Task" [ 956.595149] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.602984] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.699406] env[62521]: DEBUG nova.scheduler.client.report [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.739657] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318852, 'name': Rename_Task, 'duration_secs': 0.174535} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.740061] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.740417] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0eaa31f4-9af4-49ed-a1bf-de9a18acbd71 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.746747] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 956.746747] env[62521]: value = "task-1318856" [ 956.746747] env[62521]: _type = "Task" [ 956.746747] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.757522] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318856, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.763150] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318853, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.999808] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318854, 'name': ReconfigVM_Task, 'duration_secs': 0.235673} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.000803] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c67b9fe-cbdb-47cb-b75d-0d348b9e68ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.020255] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:36:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='e986ebeb-2013-4e50-9994-d8d533ed7ed2',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-778228551',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.020392] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.020629] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.021813] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.021813] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.021813] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.021813] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.021813] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.021813] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.021813] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.022228] env[62521]: DEBUG nova.virt.hardware [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.023274] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da6d9ab7-7388-4401-888a-5456d7ee5028 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.029062] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 957.029062] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52defd87-ec95-ac0f-f8be-0cb725fec771" [ 957.029062] env[62521]: _type = "Task" [ 957.029062] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.037923] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52defd87-ec95-ac0f-f8be-0cb725fec771, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.108994] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318855, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.205154] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.874s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.207828] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.356s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.208138] env[62521]: DEBUG nova.objects.instance [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lazy-loading 'resources' on Instance uuid ac420a77-2160-45d6-b313-c5304536c39d {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.230620] env[62521]: INFO nova.scheduler.client.report [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleted allocations for instance 563ad3b2-e53e-4682-9cde-bd6f709718de [ 957.262530] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318856, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.266032] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318853, 'name': MoveVirtualDisk_Task} progress is 29%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.355763] env[62521]: DEBUG nova.compute.manager [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 957.383395] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.383672] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.383837] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.384044] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.384208] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.384363] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.384631] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.384804] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.384978] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.385162] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.385338] env[62521]: DEBUG nova.virt.hardware [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.386316] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea72514-a39a-40ea-b2e2-3f5e373bdf23 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.395088] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f57f05-a257-428c-9147-03df6f1ab655 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.539630] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52defd87-ec95-ac0f-f8be-0cb725fec771, 'name': SearchDatastore_Task, 'duration_secs': 0.044617} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.546128] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Reconfiguring VM instance instance-0000004f to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 957.546635] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49f9a8e7-7e4b-4850-a4c9-e464b26a66c8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.567154] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 957.567154] env[62521]: value = "task-1318857" [ 957.567154] env[62521]: _type = "Task" [ 957.567154] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.576171] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318857, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.606722] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318855, 'name': ReconfigVM_Task, 'duration_secs': 0.57884} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.607106] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c/02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.607333] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance '02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c' progress to 50 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 957.726667] env[62521]: DEBUG nova.compute.manager [req-8d2a49d5-5568-41f2-b1e0-bf58e125c6d0 req-26dde9f5-1d9c-4065-96aa-d66a439b973f service nova] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Received event network-vif-plugged-fb88c8dc-e610-4e05-ba90-1a6dcf6ae085 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.727174] env[62521]: DEBUG oslo_concurrency.lockutils [req-8d2a49d5-5568-41f2-b1e0-bf58e125c6d0 req-26dde9f5-1d9c-4065-96aa-d66a439b973f service nova] Acquiring lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.728714] env[62521]: DEBUG oslo_concurrency.lockutils [req-8d2a49d5-5568-41f2-b1e0-bf58e125c6d0 req-26dde9f5-1d9c-4065-96aa-d66a439b973f service nova] Lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.728714] env[62521]: DEBUG oslo_concurrency.lockutils [req-8d2a49d5-5568-41f2-b1e0-bf58e125c6d0 req-26dde9f5-1d9c-4065-96aa-d66a439b973f service nova] Lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.729085] env[62521]: DEBUG nova.compute.manager [req-8d2a49d5-5568-41f2-b1e0-bf58e125c6d0 req-26dde9f5-1d9c-4065-96aa-d66a439b973f service nova] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] No waiting events found dispatching network-vif-plugged-fb88c8dc-e610-4e05-ba90-1a6dcf6ae085 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 957.733444] env[62521]: WARNING nova.compute.manager [req-8d2a49d5-5568-41f2-b1e0-bf58e125c6d0 req-26dde9f5-1d9c-4065-96aa-d66a439b973f service nova] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Received unexpected event network-vif-plugged-fb88c8dc-e610-4e05-ba90-1a6dcf6ae085 for instance with vm_state building and task_state spawning. [ 957.742325] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8cf76d67-de36-42e8-8215-69fa314b856c tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "563ad3b2-e53e-4682-9cde-bd6f709718de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.272s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.765204] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318856, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.772753] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318853, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.848351] env[62521]: DEBUG nova.network.neutron [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Successfully updated port: fb88c8dc-e610-4e05-ba90-1a6dcf6ae085 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 957.955287] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad59f46-7ef4-4a6a-99bb-978e9ef34ecd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.963877] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef70539c-af3f-4aa5-8b06-02a084754668 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.997233] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c0f60b-4f40-4eb7-8f97-cddbc44a1e89 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.006098] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19cdec9f-0505-4dc4-a12f-1d949b368bf3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.022017] env[62521]: DEBUG nova.compute.provider_tree [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.078318] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318857, 'name': ReconfigVM_Task, 'duration_secs': 0.264419} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.078615] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Reconfigured VM instance instance-0000004f to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 958.080091] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19007a00-b79e-4df8-85ea-18c1c7e77134 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.103200] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 17876bcc-3a7b-47b0-afd9-1215e24864ae/17876bcc-3a7b-47b0-afd9-1215e24864ae.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 958.103936] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4efd3220-6221-4e00-93d0-67ce443cd85e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.120250] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ccd540-6d43-49b9-85bb-03fc8ff5b453 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.142937] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6540c35a-79a0-4c7f-813c-914e4b9b57b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.146255] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 958.146255] env[62521]: value = "task-1318858" [ 958.146255] env[62521]: _type = "Task" [ 958.146255] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.164720] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance '02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c' progress to 67 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 958.174265] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318858, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.270698] env[62521]: DEBUG oslo_vmware.api [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318856, 'name': PowerOnVM_Task, 'duration_secs': 1.16129} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.274467] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 958.277516] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318853, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.355309] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "refresh_cache-0743a5b6-87fe-4612-a5a1-5d6527b69eb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.355309] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "refresh_cache-0743a5b6-87fe-4612-a5a1-5d6527b69eb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.355309] env[62521]: DEBUG nova.network.neutron [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.393234] env[62521]: DEBUG nova.compute.manager [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.394368] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527db9e5-49ed-4263-a68b-76a9c8f7cf9e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.525552] env[62521]: DEBUG nova.scheduler.client.report [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.660069] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318858, 'name': ReconfigVM_Task, 'duration_secs': 0.402274} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.660444] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 17876bcc-3a7b-47b0-afd9-1215e24864ae/17876bcc-3a7b-47b0-afd9-1215e24864ae.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.661686] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98966d30-dff9-4dd2-b3ce-a8320991e862 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.689517] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82fab15-7610-40cc-849b-c8c854481e73 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.711599] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443a19be-da59-4747-ae12-ecb7a81e1b66 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.730724] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6689c117-1537-444b-8121-754ad45aca78 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.738229] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.738534] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29caab85-39e6-4c7e-9606-041af6fe3a70 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.745171] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 958.745171] env[62521]: value = "task-1318859" [ 958.745171] env[62521]: _type = "Task" [ 958.745171] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.753582] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318859, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.754507] env[62521]: DEBUG nova.network.neutron [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Port 34cee820-50a0-47bd-a0c7-25f1e01fa639 binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 958.770875] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318853, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.889371] env[62521]: DEBUG nova.network.neutron [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 958.915158] env[62521]: DEBUG oslo_concurrency.lockutils [None req-286a6126-86ad-42d3-8919-2059f8cdc3d9 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 32.378s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.030681] env[62521]: DEBUG nova.network.neutron [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Updating instance_info_cache with network_info: [{"id": "fb88c8dc-e610-4e05-ba90-1a6dcf6ae085", "address": "fa:16:3e:70:13:4e", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb88c8dc-e6", "ovs_interfaceid": "fb88c8dc-e610-4e05-ba90-1a6dcf6ae085", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.032554] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.825s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.035081] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.500s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.036681] env[62521]: INFO nova.compute.claims [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 959.057540] env[62521]: INFO nova.scheduler.client.report [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted allocations for instance ac420a77-2160-45d6-b313-c5304536c39d [ 959.255784] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318859, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.270290] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318853, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.689912} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.270584] env[62521]: INFO nova.virt.vmwareapi.ds_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601/OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601.vmdk to [datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313/2919854a-4a07-48e6-87af-29f13a5c2313.vmdk. [ 959.270781] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Cleaning up location [datastore1] OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601 {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 959.270947] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_0beda93d-0fcc-4e70-894d-7dc39c3a9601 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.271222] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa495f2b-2822-43eb-a592-b70931f44f6d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.278524] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 959.278524] env[62521]: value = "task-1318860" [ 959.278524] env[62521]: _type = "Task" [ 959.278524] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.287681] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318860, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.450537] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.451356] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.535787] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "refresh_cache-0743a5b6-87fe-4612-a5a1-5d6527b69eb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.536154] env[62521]: DEBUG nova.compute.manager [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Instance network_info: |[{"id": "fb88c8dc-e610-4e05-ba90-1a6dcf6ae085", "address": "fa:16:3e:70:13:4e", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb88c8dc-e6", "ovs_interfaceid": "fb88c8dc-e610-4e05-ba90-1a6dcf6ae085", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 959.536993] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:13:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '304be4f7-4e36-4468-9ef4-e457341cef18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fb88c8dc-e610-4e05-ba90-1a6dcf6ae085', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.544995] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Creating folder: Project (f0dfb5f5bd4646ec95b7c6a2d0434175). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 959.549089] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e5530e8-895c-4f4d-95b3-88b82bee88f2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.560858] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Created folder: Project (f0dfb5f5bd4646ec95b7c6a2d0434175) in parent group-v282025. [ 959.561316] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Creating folder: Instances. Parent ref: group-v282147. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 959.563437] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d7c7ff27-e7eb-41db-922a-ba359e54fb37 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.566139] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac3ea24e-6945-4b9c-868d-7630fc080e55 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "ac420a77-2160-45d6-b313-c5304536c39d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.299s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.572984] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Created folder: Instances in parent group-v282147. [ 959.573239] env[62521]: DEBUG oslo.service.loopingcall [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.573680] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.573896] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8fc38aee-1d94-42b7-9d21-90a7913a25c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.594695] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.594695] env[62521]: value = "task-1318863" [ 959.594695] env[62521]: _type = "Task" [ 959.594695] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.603071] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318863, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.757785] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318859, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.759138] env[62521]: DEBUG nova.compute.manager [req-a1039fca-890f-4d21-99ca-a1dc47f802d4 req-0f321c95-a299-43c6-ac7c-cb230c0ace0e service nova] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Received event network-changed-fb88c8dc-e610-4e05-ba90-1a6dcf6ae085 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.759366] env[62521]: DEBUG nova.compute.manager [req-a1039fca-890f-4d21-99ca-a1dc47f802d4 req-0f321c95-a299-43c6-ac7c-cb230c0ace0e service nova] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Refreshing instance network info cache due to event network-changed-fb88c8dc-e610-4e05-ba90-1a6dcf6ae085. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 959.759617] env[62521]: DEBUG oslo_concurrency.lockutils [req-a1039fca-890f-4d21-99ca-a1dc47f802d4 req-0f321c95-a299-43c6-ac7c-cb230c0ace0e service nova] Acquiring lock "refresh_cache-0743a5b6-87fe-4612-a5a1-5d6527b69eb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.759771] env[62521]: DEBUG oslo_concurrency.lockutils [req-a1039fca-890f-4d21-99ca-a1dc47f802d4 req-0f321c95-a299-43c6-ac7c-cb230c0ace0e service nova] Acquired lock "refresh_cache-0743a5b6-87fe-4612-a5a1-5d6527b69eb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.759937] env[62521]: DEBUG nova.network.neutron [req-a1039fca-890f-4d21-99ca-a1dc47f802d4 req-0f321c95-a299-43c6-ac7c-cb230c0ace0e service nova] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Refreshing network info cache for port fb88c8dc-e610-4e05-ba90-1a6dcf6ae085 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 959.780958] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.781237] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.781425] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.793170] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318860, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038199} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.793389] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.793526] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313/2919854a-4a07-48e6-87af-29f13a5c2313.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.793789] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313/2919854a-4a07-48e6-87af-29f13a5c2313.vmdk to [datastore1] bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0/bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.794303] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c57f6088-8d68-4deb-9421-49e70177bb6d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.802811] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 959.802811] env[62521]: value = "task-1318864" [ 959.802811] env[62521]: _type = "Task" [ 959.802811] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.812143] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318864, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.954148] env[62521]: INFO nova.compute.manager [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Detaching volume 5e701085-e961-46c7-8528-8e06b1899590 [ 960.002068] env[62521]: INFO nova.virt.block_device [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Attempting to driver detach volume 5e701085-e961-46c7-8528-8e06b1899590 from mountpoint /dev/sdb [ 960.002510] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 960.002844] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282109', 'volume_id': '5e701085-e961-46c7-8528-8e06b1899590', 'name': 'volume-5e701085-e961-46c7-8528-8e06b1899590', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af91de53-2ef8-48d1-b8c1-7bd0809f4ae4', 'attached_at': '', 'detached_at': '', 'volume_id': '5e701085-e961-46c7-8528-8e06b1899590', 'serial': '5e701085-e961-46c7-8528-8e06b1899590'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 960.004305] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b511d6bc-d138-4e3e-abd3-41b961ca27ee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.027543] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9972e08-2db1-4cd5-b34b-bddc9090384b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.034953] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40fd351e-93a7-48e0-bb2f-e62d557f9dd8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.059517] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854da550-1d13-412e-94d9-52bca8b4020a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.081102] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] The volume has not been displaced from its original location: [datastore2] volume-5e701085-e961-46c7-8528-8e06b1899590/volume-5e701085-e961-46c7-8528-8e06b1899590.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 960.087032] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Reconfiguring VM instance instance-0000003a to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 960.090395] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a7e009d-d0da-45ce-a0d0-8d82923cd4cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.113413] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318863, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.115021] env[62521]: DEBUG oslo_vmware.api [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 960.115021] env[62521]: value = "task-1318865" [ 960.115021] env[62521]: _type = "Task" [ 960.115021] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.126324] env[62521]: DEBUG oslo_vmware.api [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318865, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.260573] env[62521]: DEBUG oslo_vmware.api [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318859, 'name': PowerOnVM_Task, 'duration_secs': 1.181875} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.260959] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 960.299513] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea26c46-b339-41be-896f-a1b71133c72d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.310786] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa6aad7-db14-4397-83e4-e0df0b5ffd72 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.318101] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318864, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.349321] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38cd8abd-c15f-4a56-bf84-910ac20d641d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.357844] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54cc7b2f-db60-4405-bd02-b37a6f8906d1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.373304] env[62521]: DEBUG nova.compute.provider_tree [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.431296] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.431817] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.618700] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318863, 'name': CreateVM_Task, 'duration_secs': 0.522599} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.623509] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 960.624923] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.625218] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.625662] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 960.629467] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-340e5a36-84c5-42b7-a36c-8db1f707465f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.637666] env[62521]: DEBUG oslo_vmware.api [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318865, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.641852] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 960.641852] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522e7376-c596-3527-8709-69a9014264b5" [ 960.641852] env[62521]: _type = "Task" [ 960.641852] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.651059] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522e7376-c596-3527-8709-69a9014264b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.752403] env[62521]: DEBUG nova.network.neutron [req-a1039fca-890f-4d21-99ca-a1dc47f802d4 req-0f321c95-a299-43c6-ac7c-cb230c0ace0e service nova] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Updated VIF entry in instance network info cache for port fb88c8dc-e610-4e05-ba90-1a6dcf6ae085. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 960.753318] env[62521]: DEBUG nova.network.neutron [req-a1039fca-890f-4d21-99ca-a1dc47f802d4 req-0f321c95-a299-43c6-ac7c-cb230c0ace0e service nova] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Updating instance_info_cache with network_info: [{"id": "fb88c8dc-e610-4e05-ba90-1a6dcf6ae085", "address": "fa:16:3e:70:13:4e", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb88c8dc-e6", "ovs_interfaceid": "fb88c8dc-e610-4e05-ba90-1a6dcf6ae085", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.814419] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318864, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.822269] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.822478] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.822667] env[62521]: DEBUG nova.network.neutron [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.877487] env[62521]: DEBUG nova.scheduler.client.report [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.934778] env[62521]: DEBUG nova.compute.manager [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 961.129070] env[62521]: DEBUG oslo_vmware.api [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318865, 'name': ReconfigVM_Task, 'duration_secs': 0.93317} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.129421] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Reconfigured VM instance instance-0000003a to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 961.134332] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a37ce8c-bfba-427c-9622-977451ce403c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.154585] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522e7376-c596-3527-8709-69a9014264b5, 'name': SearchDatastore_Task, 'duration_secs': 0.013058} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.156229] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.156630] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 961.156927] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.157110] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.157313] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.157703] env[62521]: DEBUG oslo_vmware.api [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 961.157703] env[62521]: value = "task-1318866" [ 961.157703] env[62521]: _type = "Task" [ 961.157703] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.158158] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-164b4a11-38c8-45d9-9439-7f3c231b31d1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.170017] env[62521]: DEBUG oslo_vmware.api [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318866, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.171445] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.171658] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 961.175080] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4489b10f-1ca1-4134-af38-162d504e94ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.176453] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.176771] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.177041] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.177297] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.177485] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.180802] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 961.180802] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a3a6eb-de07-2eda-4623-649c265974b3" [ 961.180802] env[62521]: _type = "Task" [ 961.180802] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.181457] env[62521]: INFO nova.compute.manager [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Terminating instance [ 961.188262] env[62521]: DEBUG nova.compute.manager [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 961.188399] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.189351] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29afae64-c8d5-405e-8028-47054680c0d0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.198156] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a3a6eb-de07-2eda-4623-649c265974b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.200539] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 961.200816] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43a69e0d-078c-4f59-b777-defc6c1705b6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.207495] env[62521]: DEBUG oslo_vmware.api [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 961.207495] env[62521]: value = "task-1318867" [ 961.207495] env[62521]: _type = "Task" [ 961.207495] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.218634] env[62521]: DEBUG oslo_vmware.api [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.256166] env[62521]: DEBUG oslo_concurrency.lockutils [req-a1039fca-890f-4d21-99ca-a1dc47f802d4 req-0f321c95-a299-43c6-ac7c-cb230c0ace0e service nova] Releasing lock "refresh_cache-0743a5b6-87fe-4612-a5a1-5d6527b69eb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.277539] env[62521]: INFO nova.compute.manager [None req-5d2d1024-3282-4c22-b578-209f873ec70c tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance to original state: 'active' [ 961.315936] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318864, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.384225] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.349s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.385025] env[62521]: DEBUG nova.compute.manager [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 961.462053] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.463044] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.464831] env[62521]: INFO nova.compute.claims [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.587655] env[62521]: DEBUG nova.network.neutron [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance_info_cache with network_info: [{"id": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "address": "fa:16:3e:8c:86:c8", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34cee820-50", "ovs_interfaceid": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.672123] env[62521]: DEBUG oslo_vmware.api [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318866, 'name': ReconfigVM_Task, 'duration_secs': 0.429538} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.672123] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282109', 'volume_id': '5e701085-e961-46c7-8528-8e06b1899590', 'name': 'volume-5e701085-e961-46c7-8528-8e06b1899590', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af91de53-2ef8-48d1-b8c1-7bd0809f4ae4', 'attached_at': '', 'detached_at': '', 'volume_id': '5e701085-e961-46c7-8528-8e06b1899590', 'serial': '5e701085-e961-46c7-8528-8e06b1899590'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 961.693586] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a3a6eb-de07-2eda-4623-649c265974b3, 'name': SearchDatastore_Task, 'duration_secs': 0.019563} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.694563] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f63bc3-fb37-4235-bfdb-6907500665a8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.700622] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 961.700622] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d95889-35dd-bdec-3ede-e43b00c8c96a" [ 961.700622] env[62521]: _type = "Task" [ 961.700622] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.709751] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d95889-35dd-bdec-3ede-e43b00c8c96a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.718154] env[62521]: DEBUG oslo_vmware.api [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318867, 'name': PowerOffVM_Task, 'duration_secs': 0.254482} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.718270] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 961.718442] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 961.718711] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e26bce08-dbee-47d2-99a7-f376cf12f423 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.785945] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 961.786899] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 961.786899] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Deleting the datastore file [datastore2] 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.789492] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1581b62f-9a2a-419e-994c-c3b59265ae95 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.796632] env[62521]: DEBUG oslo_vmware.api [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 961.796632] env[62521]: value = "task-1318869" [ 961.796632] env[62521]: _type = "Task" [ 961.796632] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.806322] env[62521]: DEBUG oslo_vmware.api [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318869, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.813636] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318864, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.892520] env[62521]: DEBUG nova.compute.utils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 961.894484] env[62521]: DEBUG nova.compute.manager [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 961.894764] env[62521]: DEBUG nova.network.neutron [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 961.955955] env[62521]: DEBUG nova.policy [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ce6cad0be5f455eab09b1d348268329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '961924204a584b61a4a85a761821910d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 962.094187] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.213079] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d95889-35dd-bdec-3ede-e43b00c8c96a, 'name': SearchDatastore_Task, 'duration_secs': 0.01413} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.213364] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.213647] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 0743a5b6-87fe-4612-a5a1-5d6527b69eb2/0743a5b6-87fe-4612-a5a1-5d6527b69eb2.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 962.213986] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-46748259-cdb7-49e5-8d75-5b68907c6db6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.220537] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 962.220537] env[62521]: value = "task-1318870" [ 962.220537] env[62521]: _type = "Task" [ 962.220537] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.224965] env[62521]: DEBUG nova.objects.instance [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lazy-loading 'flavor' on Instance uuid af91de53-2ef8-48d1-b8c1-7bd0809f4ae4 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.231607] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318870, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.301984] env[62521]: DEBUG nova.network.neutron [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Successfully created port: 23b0cfbc-ec82-4a35-a248-249e86ecec84 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 962.308564] env[62521]: DEBUG oslo_vmware.api [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318869, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.387465} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.312691] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.312691] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 962.313671] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 962.313917] env[62521]: INFO nova.compute.manager [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Took 1.13 seconds to destroy the instance on the hypervisor. [ 962.314189] env[62521]: DEBUG oslo.service.loopingcall [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.314929] env[62521]: DEBUG nova.compute.manager [-] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.315040] env[62521]: DEBUG nova.network.neutron [-] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 962.321910] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318864, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.374117} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.322504] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2919854a-4a07-48e6-87af-29f13a5c2313/2919854a-4a07-48e6-87af-29f13a5c2313.vmdk to [datastore1] bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0/bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 962.323616] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3552402-9914-4c78-b1fc-486dd1497ef3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.347044] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0/bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0.vmdk or device None with type streamOptimized {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.349534] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10469e9a-d4d8-4161-8fcf-7cb4e73f8a43 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.370324] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 962.370324] env[62521]: value = "task-1318871" [ 962.370324] env[62521]: _type = "Task" [ 962.370324] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.378142] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318871, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.398433] env[62521]: DEBUG nova.compute.manager [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 962.405198] env[62521]: DEBUG oslo_concurrency.lockutils [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "17876bcc-3a7b-47b0-afd9-1215e24864ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.405492] env[62521]: DEBUG oslo_concurrency.lockutils [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "17876bcc-3a7b-47b0-afd9-1215e24864ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.405717] env[62521]: DEBUG oslo_concurrency.lockutils [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "17876bcc-3a7b-47b0-afd9-1215e24864ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.405911] env[62521]: DEBUG oslo_concurrency.lockutils [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "17876bcc-3a7b-47b0-afd9-1215e24864ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.406098] env[62521]: DEBUG oslo_concurrency.lockutils [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "17876bcc-3a7b-47b0-afd9-1215e24864ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.408325] env[62521]: INFO nova.compute.manager [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Terminating instance [ 962.410181] env[62521]: DEBUG nova.compute.manager [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 962.410381] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 962.411255] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3ee948-fe53-4fc0-b1e6-dff41bcf1e29 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.418868] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 962.419129] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94a4576d-6f71-4409-a408-ee0c0e8fe111 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.425029] env[62521]: DEBUG oslo_vmware.api [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 962.425029] env[62521]: value = "task-1318872" [ 962.425029] env[62521]: _type = "Task" [ 962.425029] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.433448] env[62521]: DEBUG oslo_vmware.api [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318872, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.617553] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b72d3d5-3056-4e8b-b3b4-b83de2843d2e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.644188] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40478ba-e44b-4d0d-bb36-3e2245c8d07a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.651863] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance '02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c' progress to 83 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 962.726666] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56fa21dd-ec96-49ed-b56e-9df808a2841c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.741742] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318870, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.746078] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5528f3-633f-425c-baa0-ee5cf1866f42 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.786133] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f98572f-e356-467e-aecf-0a27e91c8589 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.796053] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64238f95-8ca3-4a20-886d-c9d99bfffbb7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.810978] env[62521]: DEBUG nova.compute.provider_tree [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 962.882291] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.938285] env[62521]: DEBUG oslo_vmware.api [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318872, 'name': PowerOffVM_Task, 'duration_secs': 0.32543} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.938733] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 962.938862] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 962.939174] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2973b65c-f264-4618-969b-6c87aa5652c6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.003534] env[62521]: DEBUG nova.compute.manager [req-01ea0c03-a202-4951-8bb0-d6844f415459 req-5dd242ad-9a6a-4f92-b97e-d690c168ed72 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Received event network-vif-deleted-06aa91ed-3b1d-4c27-a91f-6246e75add4e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.003790] env[62521]: INFO nova.compute.manager [req-01ea0c03-a202-4951-8bb0-d6844f415459 req-5dd242ad-9a6a-4f92-b97e-d690c168ed72 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Neutron deleted interface 06aa91ed-3b1d-4c27-a91f-6246e75add4e; detaching it from the instance and deleting it from the info cache [ 963.003970] env[62521]: DEBUG nova.network.neutron [req-01ea0c03-a202-4951-8bb0-d6844f415459 req-5dd242ad-9a6a-4f92-b97e-d690c168ed72 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.011691] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 963.011979] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 963.012196] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Deleting the datastore file [datastore1] 17876bcc-3a7b-47b0-afd9-1215e24864ae {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 963.012721] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09bcca14-bd13-41f9-a493-36012c7fc621 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.020057] env[62521]: DEBUG oslo_vmware.api [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 963.020057] env[62521]: value = "task-1318874" [ 963.020057] env[62521]: _type = "Task" [ 963.020057] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.030620] env[62521]: DEBUG oslo_vmware.api [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318874, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.163033] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.164047] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c959a4f-37f6-45a6-8424-4210c706ff5e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.171915] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 963.171915] env[62521]: value = "task-1318875" [ 963.171915] env[62521]: _type = "Task" [ 963.171915] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.181417] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318875, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.231335] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318870, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.775215} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.231660] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 0743a5b6-87fe-4612-a5a1-5d6527b69eb2/0743a5b6-87fe-4612-a5a1-5d6527b69eb2.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.231903] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.232203] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ee9b770-51a6-4907-b16e-4d275c182efe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.239631] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 963.239631] env[62521]: value = "task-1318876" [ 963.239631] env[62521]: _type = "Task" [ 963.239631] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.243068] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f90c23b2-241c-46f7-899e-f3774c69bead tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.792s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.251817] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318876, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.334976] env[62521]: ERROR nova.scheduler.client.report [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [req-b1f90aa7-cf25-4a9e-bc29-fdf4818a2ee5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b1f90aa7-cf25-4a9e-bc29-fdf4818a2ee5"}]} [ 963.351573] env[62521]: DEBUG nova.scheduler.client.report [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 963.366565] env[62521]: DEBUG nova.scheduler.client.report [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 963.366870] env[62521]: DEBUG nova.compute.provider_tree [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 963.378969] env[62521]: DEBUG nova.scheduler.client.report [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 963.387514] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318871, 'name': ReconfigVM_Task, 'duration_secs': 0.610502} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.387844] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Reconfigured VM instance instance-00000041 to attach disk [datastore1] bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0/bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0.vmdk or device None with type streamOptimized {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.388484] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a7961f0-4962-4d0c-a7a3-9c486e8d5b90 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.394874] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 963.394874] env[62521]: value = "task-1318877" [ 963.394874] env[62521]: _type = "Task" [ 963.394874] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.399680] env[62521]: DEBUG nova.scheduler.client.report [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 963.409447] env[62521]: DEBUG nova.compute.manager [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 963.411303] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318877, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.431007] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 963.431284] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 963.431555] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.431812] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 963.431949] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.432126] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 963.432346] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 963.432512] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 963.432672] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 963.432837] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 963.433125] env[62521]: DEBUG nova.virt.hardware [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.433922] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f94626-8a77-4b82-8783-ab6f4cb6f013 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.445714] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef7ff47-81e4-4043-a2f5-b022b58313c3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.464526] env[62521]: DEBUG nova.network.neutron [-] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.506235] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a03b9f78-5131-49dc-b2a0-a71fb35c2ae0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.521209] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885c96da-8b62-4402-95b8-4e262c326eec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.542582] env[62521]: DEBUG oslo_vmware.api [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318874, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190628} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.542985] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.543339] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 963.543504] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 963.543739] env[62521]: INFO nova.compute.manager [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Took 1.13 seconds to destroy the instance on the hypervisor. [ 963.544049] env[62521]: DEBUG oslo.service.loopingcall [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.544316] env[62521]: DEBUG nova.compute.manager [-] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 963.544461] env[62521]: DEBUG nova.network.neutron [-] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 963.558257] env[62521]: DEBUG nova.compute.manager [req-01ea0c03-a202-4951-8bb0-d6844f415459 req-5dd242ad-9a6a-4f92-b97e-d690c168ed72 service nova] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Detach interface failed, port_id=06aa91ed-3b1d-4c27-a91f-6246e75add4e, reason: Instance 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 963.639604] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3203fea4-34af-4ff3-b0e2-0b1e2fcfb837 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.647519] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cae0151-b8e2-4078-80c2-8d8a987e14ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.683495] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e90be8-f717-4db4-aeaf-10124b0c4353 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.691885] env[62521]: DEBUG oslo_vmware.api [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318875, 'name': PowerOnVM_Task, 'duration_secs': 0.46728} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.694114] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 963.694328] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8ddc4e-9c4b-43de-8919-a20793bd59fb tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance '02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c' progress to 100 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 963.698937] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683d1e59-166d-4183-8174-842a6ff3652f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.715443] env[62521]: DEBUG nova.compute.provider_tree [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.733068] env[62521]: DEBUG nova.compute.manager [req-6f7ed7ca-504d-42de-87e3-9abc02b44752 req-9edde7fb-d1f0-43ba-80cd-5689bd342013 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Received event network-vif-plugged-23b0cfbc-ec82-4a35-a248-249e86ecec84 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.733332] env[62521]: DEBUG oslo_concurrency.lockutils [req-6f7ed7ca-504d-42de-87e3-9abc02b44752 req-9edde7fb-d1f0-43ba-80cd-5689bd342013 service nova] Acquiring lock "b8158051-98b6-47d8-84c5-d1d56515afe8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.733540] env[62521]: DEBUG oslo_concurrency.lockutils [req-6f7ed7ca-504d-42de-87e3-9abc02b44752 req-9edde7fb-d1f0-43ba-80cd-5689bd342013 service nova] Lock "b8158051-98b6-47d8-84c5-d1d56515afe8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.733671] env[62521]: DEBUG oslo_concurrency.lockutils [req-6f7ed7ca-504d-42de-87e3-9abc02b44752 req-9edde7fb-d1f0-43ba-80cd-5689bd342013 service nova] Lock "b8158051-98b6-47d8-84c5-d1d56515afe8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.733845] env[62521]: DEBUG nova.compute.manager [req-6f7ed7ca-504d-42de-87e3-9abc02b44752 req-9edde7fb-d1f0-43ba-80cd-5689bd342013 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] No waiting events found dispatching network-vif-plugged-23b0cfbc-ec82-4a35-a248-249e86ecec84 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.734018] env[62521]: WARNING nova.compute.manager [req-6f7ed7ca-504d-42de-87e3-9abc02b44752 req-9edde7fb-d1f0-43ba-80cd-5689bd342013 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Received unexpected event network-vif-plugged-23b0cfbc-ec82-4a35-a248-249e86ecec84 for instance with vm_state building and task_state spawning. [ 963.750660] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318876, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070966} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.750818] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.751636] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513072ea-79f7-4a63-8ec4-1b2cbd3aab2c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.774533] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 0743a5b6-87fe-4612-a5a1-5d6527b69eb2/0743a5b6-87fe-4612-a5a1-5d6527b69eb2.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.775917] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8e65764-e21c-4c00-a861-fe2e7687e246 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.799471] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 963.799471] env[62521]: value = "task-1318878" [ 963.799471] env[62521]: _type = "Task" [ 963.799471] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.807023] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318878, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.834376] env[62521]: DEBUG nova.network.neutron [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Successfully updated port: 23b0cfbc-ec82-4a35-a248-249e86ecec84 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.864035] env[62521]: DEBUG oslo_vmware.rw_handles [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52399f2a-09cc-82b5-feec-8e2f0ca9c197/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 963.865138] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d25d45-75be-412a-ad03-14466295cdd3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.871930] env[62521]: DEBUG oslo_vmware.rw_handles [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52399f2a-09cc-82b5-feec-8e2f0ca9c197/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 963.872126] env[62521]: ERROR oslo_vmware.rw_handles [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52399f2a-09cc-82b5-feec-8e2f0ca9c197/disk-0.vmdk due to incomplete transfer. [ 963.872361] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-826da6ab-11de-4503-a692-3ac075546d71 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.878659] env[62521]: DEBUG oslo_vmware.rw_handles [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52399f2a-09cc-82b5-feec-8e2f0ca9c197/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 963.878867] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Uploaded image 07530e24-0ad3-44e9-9fb4-ee46d23c744c to the Glance image server {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 963.881193] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 963.881807] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.882058] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.882282] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.882481] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.882653] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.884147] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ffcb5f7f-7fa7-4f59-8b15-48e47d266b69 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.885925] env[62521]: INFO nova.compute.manager [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Terminating instance [ 963.888180] env[62521]: DEBUG nova.compute.manager [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 963.888390] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 963.889155] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cdca744-bcd1-430f-a464-f462915d1618 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.894984] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 963.894984] env[62521]: value = "task-1318879" [ 963.894984] env[62521]: _type = "Task" [ 963.894984] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.897239] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.902634] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8dc08a89-4773-4a67-bdc5-c446998b4f1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.908639] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318879, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.912707] env[62521]: DEBUG oslo_vmware.api [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 963.912707] env[62521]: value = "task-1318880" [ 963.912707] env[62521]: _type = "Task" [ 963.912707] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.913024] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318877, 'name': Rename_Task, 'duration_secs': 0.232357} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.913361] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.917032] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62deea8b-5cd0-40c0-8d10-86f222680b2b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.924106] env[62521]: DEBUG oslo_vmware.api [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.925718] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 963.925718] env[62521]: value = "task-1318881" [ 963.925718] env[62521]: _type = "Task" [ 963.925718] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.933568] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318881, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.967413] env[62521]: INFO nova.compute.manager [-] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Took 1.65 seconds to deallocate network for instance. [ 964.218247] env[62521]: DEBUG nova.scheduler.client.report [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.293072] env[62521]: DEBUG nova.network.neutron [-] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.312380] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318878, 'name': ReconfigVM_Task, 'duration_secs': 0.350902} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.312672] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 0743a5b6-87fe-4612-a5a1-5d6527b69eb2/0743a5b6-87fe-4612-a5a1-5d6527b69eb2.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.313297] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-746f493b-2b8d-43ba-8980-844600931aa0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.319848] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 964.319848] env[62521]: value = "task-1318882" [ 964.319848] env[62521]: _type = "Task" [ 964.319848] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.328559] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318882, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.342096] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.342096] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.342096] env[62521]: DEBUG nova.network.neutron [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.406939] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318879, 'name': Destroy_Task, 'duration_secs': 0.355869} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.407258] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Destroyed the VM [ 964.407513] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 964.407802] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d3b6711f-c7ea-44e7-b6ee-97bb9922d4f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.414135] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 964.414135] env[62521]: value = "task-1318883" [ 964.414135] env[62521]: _type = "Task" [ 964.414135] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.424144] env[62521]: DEBUG oslo_vmware.api [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318880, 'name': PowerOffVM_Task, 'duration_secs': 0.291913} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.426976] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.427591] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 964.427591] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318883, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.427730] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1f57ffa-b7a2-4416-ac6c-68779b253e16 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.436425] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318881, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.474066] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.503073] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.503315] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.503501] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleting the datastore file [datastore1] af91de53-2ef8-48d1-b8c1-7bd0809f4ae4 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.503773] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ce8c128-9288-450d-aaf3-342c3de9f782 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.513982] env[62521]: DEBUG oslo_vmware.api [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 964.513982] env[62521]: value = "task-1318885" [ 964.513982] env[62521]: _type = "Task" [ 964.513982] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.522032] env[62521]: DEBUG oslo_vmware.api [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318885, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.723539] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.261s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.724275] env[62521]: DEBUG nova.compute.manager [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 964.726959] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.253s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.727288] env[62521]: DEBUG nova.objects.instance [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'resources' on Instance uuid 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.795633] env[62521]: INFO nova.compute.manager [-] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Took 1.25 seconds to deallocate network for instance. [ 964.830192] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318882, 'name': Rename_Task, 'duration_secs': 0.160165} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.830476] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 964.831320] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-deebf23d-d749-4456-af1d-a81ae01c5adc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.839405] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 964.839405] env[62521]: value = "task-1318886" [ 964.839405] env[62521]: _type = "Task" [ 964.839405] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.849608] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318886, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.875012] env[62521]: DEBUG nova.network.neutron [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 964.927665] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318883, 'name': RemoveSnapshot_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.936060] env[62521]: DEBUG oslo_vmware.api [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318881, 'name': PowerOnVM_Task, 'duration_secs': 0.917987} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.936360] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 965.013812] env[62521]: DEBUG nova.network.neutron [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updating instance_info_cache with network_info: [{"id": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "address": "fa:16:3e:37:05:5a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b0cfbc-ec", "ovs_interfaceid": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.025574] env[62521]: DEBUG oslo_vmware.api [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318885, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.466617} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.025945] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.026062] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.026249] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.026434] env[62521]: INFO nova.compute.manager [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Took 1.14 seconds to destroy the instance on the hypervisor. [ 965.026765] env[62521]: DEBUG oslo.service.loopingcall [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.026972] env[62521]: DEBUG nova.compute.manager [-] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.027080] env[62521]: DEBUG nova.network.neutron [-] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 965.030620] env[62521]: DEBUG nova.compute.manager [req-b785046f-1930-41bb-8ae1-edc4e5055d89 req-be148937-dee1-4181-ad1d-c1cf9b4427db service nova] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Received event network-vif-deleted-3b0934da-f42d-4e91-ae7c-a4f254d6b86c {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.064307] env[62521]: DEBUG nova.compute.manager [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.065814] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847ddd3c-2893-4527-b57b-4cb9d96a3c47 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.230186] env[62521]: DEBUG nova.compute.utils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 965.234255] env[62521]: DEBUG nova.compute.manager [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 965.234534] env[62521]: DEBUG nova.network.neutron [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 965.306295] env[62521]: DEBUG oslo_concurrency.lockutils [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.306295] env[62521]: DEBUG nova.policy [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76d85009c2c1416b9ea049003e6d96a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c533bd4917b4466e9c5c76a9caf8e807', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 965.354382] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318886, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.439591] env[62521]: DEBUG oslo_vmware.api [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318883, 'name': RemoveSnapshot_Task, 'duration_secs': 0.669958} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.439591] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 965.439591] env[62521]: INFO nova.compute.manager [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Took 16.07 seconds to snapshot the instance on the hypervisor. [ 965.475172] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77696b0d-6473-4047-a358-895af2df144a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.484518] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e8ec16-f333-4187-b4fb-7fd05e7b5a73 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.518481] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0cd1c3a-09d4-4c4b-8040-cb88c272ab70 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.521364] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.521675] env[62521]: DEBUG nova.compute.manager [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Instance network_info: |[{"id": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "address": "fa:16:3e:37:05:5a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b0cfbc-ec", "ovs_interfaceid": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.522098] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:05:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92cdccfd-4b10-4024-b724-5f22792dd4de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '23b0cfbc-ec82-4a35-a248-249e86ecec84', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.529446] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Creating folder: Project (961924204a584b61a4a85a761821910d). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 965.530145] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ef22300-c6af-44d2-a4ab-2641a6253287 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.535949] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db4974c-8c1b-43c5-b33e-7c4f18ff229c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.552114] env[62521]: DEBUG nova.compute.provider_tree [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.554406] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Created folder: Project (961924204a584b61a4a85a761821910d) in parent group-v282025. [ 965.554623] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Creating folder: Instances. Parent ref: group-v282150. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 965.555055] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-291d7626-8b7c-457e-b810-ce0e07d7f899 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.563668] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Created folder: Instances in parent group-v282150. [ 965.563922] env[62521]: DEBUG oslo.service.loopingcall [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.564434] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 965.564618] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6766723-db90-4aaf-8043-c8d39c23c4f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.589533] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.589533] env[62521]: value = "task-1318889" [ 965.589533] env[62521]: _type = "Task" [ 965.589533] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.594098] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48616a2d-118a-4fac-97cf-fedc06ecd9e4 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.592s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.598849] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318889, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.735560] env[62521]: DEBUG nova.compute.manager [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 965.746712] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.747133] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.747505] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 965.815290] env[62521]: DEBUG nova.network.neutron [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Successfully created port: 45c9722b-0936-4e02-8f69-17a30523490b {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.849131] env[62521]: DEBUG nova.compute.manager [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Received event network-changed-23b0cfbc-ec82-4a35-a248-249e86ecec84 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.849382] env[62521]: DEBUG nova.compute.manager [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Refreshing instance network info cache due to event network-changed-23b0cfbc-ec82-4a35-a248-249e86ecec84. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.849646] env[62521]: DEBUG oslo_concurrency.lockutils [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] Acquiring lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.849835] env[62521]: DEBUG oslo_concurrency.lockutils [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] Acquired lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.850068] env[62521]: DEBUG nova.network.neutron [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Refreshing network info cache for port 23b0cfbc-ec82-4a35-a248-249e86ecec84 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.859492] env[62521]: DEBUG oslo_vmware.api [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318886, 'name': PowerOnVM_Task, 'duration_secs': 0.545696} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.859492] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 965.859492] env[62521]: INFO nova.compute.manager [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Took 8.50 seconds to spawn the instance on the hypervisor. [ 965.859492] env[62521]: DEBUG nova.compute.manager [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.862027] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6895a72d-37e6-48b0-a5de-5f2e4cc5fee9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.014628] env[62521]: DEBUG nova.compute.manager [None req-2e214c5c-21b6-41b2-a0d0-c4932224f03f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Found 1 images (rotation: 2) {{(pid=62521) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 966.056278] env[62521]: DEBUG nova.scheduler.client.report [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.101538] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318889, 'name': CreateVM_Task, 'duration_secs': 0.39855} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.101776] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 966.102711] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.102929] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.103833] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.104171] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2484c2e-97b4-4b8a-9c04-5b14c2af7b5c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.107967] env[62521]: DEBUG nova.network.neutron [-] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.115018] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 966.115018] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a19135-d46f-d016-45e0-2c2567728bc7" [ 966.115018] env[62521]: _type = "Task" [ 966.115018] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.120981] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a19135-d46f-d016-45e0-2c2567728bc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.290389] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.290763] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.290995] env[62521]: DEBUG nova.compute.manager [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Going to confirm migration 4 {{(pid=62521) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 966.384930] env[62521]: INFO nova.compute.manager [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Took 15.57 seconds to build instance. [ 966.562744] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.564793] env[62521]: DEBUG oslo_concurrency.lockutils [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.261s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.565049] env[62521]: DEBUG oslo_concurrency.lockutils [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.588099] env[62521]: DEBUG nova.network.neutron [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updated VIF entry in instance network info cache for port 23b0cfbc-ec82-4a35-a248-249e86ecec84. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.589042] env[62521]: DEBUG nova.network.neutron [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updating instance_info_cache with network_info: [{"id": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "address": "fa:16:3e:37:05:5a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b0cfbc-ec", "ovs_interfaceid": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.598678] env[62521]: INFO nova.scheduler.client.report [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Deleted allocations for instance 538ed18e-920e-4ed4-a9e1-a6bfbae7f282 [ 966.600911] env[62521]: INFO nova.scheduler.client.report [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Deleted allocations for instance 17876bcc-3a7b-47b0-afd9-1215e24864ae [ 966.610125] env[62521]: INFO nova.compute.manager [-] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Took 1.58 seconds to deallocate network for instance. [ 966.624781] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a19135-d46f-d016-45e0-2c2567728bc7, 'name': SearchDatastore_Task, 'duration_secs': 0.014888} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.625108] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.625339] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.625574] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.625723] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.625905] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.626203] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0e363cb4-0317-439c-88d2-145d1f72afac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.639911] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.640121] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 966.640933] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf205a14-677a-41ed-9373-46af27e40b5d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.646372] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 966.646372] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523ad12c-4369-3e7b-88df-6efb76dc409f" [ 966.646372] env[62521]: _type = "Task" [ 966.646372] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.659123] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523ad12c-4369-3e7b-88df-6efb76dc409f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.747105] env[62521]: DEBUG nova.compute.manager [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 966.761088] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Skipping network cache update for instance because it is being deleted. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9977}} [ 966.771031] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.771031] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.771031] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.771031] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.771031] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.771031] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.771031] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.775033] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.775033] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.775033] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.775033] env[62521]: DEBUG nova.virt.hardware [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.775033] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26529ce5-57c5-4f65-b5d8-680a71d7ca71 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.781451] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded82ca4-a42d-4efe-a595-49f05e4c82c0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.822683] env[62521]: DEBUG nova.compute.manager [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.823829] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c289c3-20b9-4fa2-baed-dcf980bc5200 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.887552] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b498c808-9587-4a0d-b2e0-26092496d4f8 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.086s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.908127] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.908326] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.912032] env[62521]: DEBUG nova.network.neutron [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.912032] env[62521]: DEBUG nova.objects.instance [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lazy-loading 'info_cache' on Instance uuid 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.092285] env[62521]: DEBUG oslo_concurrency.lockutils [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] Releasing lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.092645] env[62521]: DEBUG nova.compute.manager [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Received event network-vif-deleted-6cf47748-b470-4bc0-8a01-23c0154e45be {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.092704] env[62521]: INFO nova.compute.manager [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Neutron deleted interface 6cf47748-b470-4bc0-8a01-23c0154e45be; detaching it from the instance and deleting it from the info cache [ 967.092881] env[62521]: DEBUG nova.network.neutron [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.121025] env[62521]: DEBUG oslo_concurrency.lockutils [None req-55913945-06f8-40b2-ad68-c9a8bba7552b tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "17876bcc-3a7b-47b0-afd9-1215e24864ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.715s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.123555] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.123806] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.124033] env[62521]: DEBUG nova.objects.instance [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lazy-loading 'resources' on Instance uuid af91de53-2ef8-48d1-b8c1-7bd0809f4ae4 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.139111] env[62521]: DEBUG oslo_concurrency.lockutils [None req-48d4a0c1-bda5-44e9-844c-97bb7f8d3f22 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "538ed18e-920e-4ed4-a9e1-a6bfbae7f282" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.962s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.159476] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523ad12c-4369-3e7b-88df-6efb76dc409f, 'name': SearchDatastore_Task, 'duration_secs': 0.01568} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.160410] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.160643] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.160868] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.161070] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.161248] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.163207] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6a43557-2c9e-4b1e-a460-e077124f5621 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.166016] env[62521]: INFO nova.compute.manager [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Terminating instance [ 967.168735] env[62521]: DEBUG nova.compute.manager [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 967.168954] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 967.170594] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb3dd41-16fc-49f9-b699-e9546b91e712 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.175163] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 967.175163] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521c0a65-c0da-c5fc-ae58-62c61dd7417a" [ 967.175163] env[62521]: _type = "Task" [ 967.175163] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.183969] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 967.183969] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0955fe9-1e4b-4515-a365-01cc53f3c4c3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.189750] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521c0a65-c0da-c5fc-ae58-62c61dd7417a, 'name': SearchDatastore_Task, 'duration_secs': 0.013289} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.191019] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.191310] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] b8158051-98b6-47d8-84c5-d1d56515afe8/b8158051-98b6-47d8-84c5-d1d56515afe8.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 967.191715] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e8537ac-aa89-4c80-b561-828ac500fdab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.194761] env[62521]: DEBUG oslo_vmware.api [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 967.194761] env[62521]: value = "task-1318890" [ 967.194761] env[62521]: _type = "Task" [ 967.194761] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.199470] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 967.199470] env[62521]: value = "task-1318891" [ 967.199470] env[62521]: _type = "Task" [ 967.199470] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.205744] env[62521]: DEBUG oslo_vmware.api [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318890, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.210668] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.295401] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.295529] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquired lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.295690] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Forcefully refreshing network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 967.340707] env[62521]: INFO nova.compute.manager [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] instance snapshotting [ 967.341489] env[62521]: DEBUG nova.objects.instance [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'flavor' on Instance uuid 11e61e72-a311-4c43-bf53-df82ad9d70f7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.554739] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7d6d73-625b-431e-8852-299b845f2c0e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.564721] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f04c5b82-2026-45ab-a6fe-c4eb4f22343a tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Suspending the VM {{(pid=62521) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 967.564721] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-69e77a56-c2ba-4593-b768-d31e5603040d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.570909] env[62521]: DEBUG oslo_vmware.api [None req-f04c5b82-2026-45ab-a6fe-c4eb4f22343a tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 967.570909] env[62521]: value = "task-1318892" [ 967.570909] env[62521]: _type = "Task" [ 967.570909] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.579597] env[62521]: DEBUG oslo_vmware.api [None req-f04c5b82-2026-45ab-a6fe-c4eb4f22343a tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318892, 'name': SuspendVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.596551] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ee29fd9-2702-47ce-a8b0-e3fb9675ab82 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.605123] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b07a163-ebf2-4b09-8744-cf84532c9991 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.637365] env[62521]: DEBUG nova.compute.manager [req-8dff97dc-95fb-401c-b70d-ec0405ab27d7 req-dfd111e8-6182-494b-9fef-d9a59ae1d655 service nova] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Detach interface failed, port_id=6cf47748-b470-4bc0-8a01-23c0154e45be, reason: Instance af91de53-2ef8-48d1-b8c1-7bd0809f4ae4 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 967.711764] env[62521]: DEBUG oslo_vmware.api [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318890, 'name': PowerOffVM_Task, 'duration_secs': 0.199012} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.716721] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 967.717323] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 967.717713] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318891, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.719304] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1581aac0-c916-4da8-93fc-4bd79c237343 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.721785] env[62521]: DEBUG nova.compute.manager [req-7c081c8a-345c-4429-8e02-8b9a7d742c15 req-d4dcf170-701f-4c71-b2bb-81bad84ec371 service nova] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Received event network-vif-plugged-45c9722b-0936-4e02-8f69-17a30523490b {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.722058] env[62521]: DEBUG oslo_concurrency.lockutils [req-7c081c8a-345c-4429-8e02-8b9a7d742c15 req-d4dcf170-701f-4c71-b2bb-81bad84ec371 service nova] Acquiring lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.722317] env[62521]: DEBUG oslo_concurrency.lockutils [req-7c081c8a-345c-4429-8e02-8b9a7d742c15 req-d4dcf170-701f-4c71-b2bb-81bad84ec371 service nova] Lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.722530] env[62521]: DEBUG oslo_concurrency.lockutils [req-7c081c8a-345c-4429-8e02-8b9a7d742c15 req-d4dcf170-701f-4c71-b2bb-81bad84ec371 service nova] Lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.722742] env[62521]: DEBUG nova.compute.manager [req-7c081c8a-345c-4429-8e02-8b9a7d742c15 req-d4dcf170-701f-4c71-b2bb-81bad84ec371 service nova] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] No waiting events found dispatching network-vif-plugged-45c9722b-0936-4e02-8f69-17a30523490b {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 967.722958] env[62521]: WARNING nova.compute.manager [req-7c081c8a-345c-4429-8e02-8b9a7d742c15 req-d4dcf170-701f-4c71-b2bb-81bad84ec371 service nova] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Received unexpected event network-vif-plugged-45c9722b-0936-4e02-8f69-17a30523490b for instance with vm_state building and task_state spawning. [ 967.784193] env[62521]: DEBUG oslo_concurrency.lockutils [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "d5f6492d-ca1b-4e74-b792-b04d55c33660" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.784499] env[62521]: DEBUG oslo_concurrency.lockutils [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.784764] env[62521]: DEBUG oslo_concurrency.lockutils [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "d5f6492d-ca1b-4e74-b792-b04d55c33660-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.784969] env[62521]: DEBUG oslo_concurrency.lockutils [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.785199] env[62521]: DEBUG oslo_concurrency.lockutils [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.788186] env[62521]: INFO nova.compute.manager [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Terminating instance [ 967.792381] env[62521]: DEBUG nova.compute.manager [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 967.792803] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 967.793667] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19134693-2d53-4b8f-9f97-72a20a212987 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.804942] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 967.806122] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 967.806122] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleting the datastore file [datastore2] 0743a5b6-87fe-4612-a5a1-5d6527b69eb2 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 967.806864] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33534684-1409-41b8-bbb4-1119ec08624e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.813636] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 967.814360] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f12a689e-4e57-486c-9a88-8d01e8ae7de7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.822024] env[62521]: DEBUG oslo_vmware.api [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 967.822024] env[62521]: value = "task-1318894" [ 967.822024] env[62521]: _type = "Task" [ 967.822024] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.824440] env[62521]: DEBUG oslo_vmware.api [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 967.824440] env[62521]: value = "task-1318895" [ 967.824440] env[62521]: _type = "Task" [ 967.824440] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.836223] env[62521]: DEBUG oslo_vmware.api [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318894, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.842059] env[62521]: DEBUG oslo_vmware.api [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318895, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.850341] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776b77a4-2f9a-4059-9a9e-341bb084d714 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.880679] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e38c2e1-6ee8-4828-8d6b-e82166071a18 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.896345] env[62521]: DEBUG nova.network.neutron [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Successfully updated port: 45c9722b-0936-4e02-8f69-17a30523490b {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.926228] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2450fef-a3e5-452d-8556-75b3b27f37fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.936016] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd40872-3204-4bbb-9578-e661c6f0bc49 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.975098] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6632df68-9071-4fbe-b419-83e2bd935624 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.983829] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ba43b9-f56a-40ec-a7e9-2e9fd2bd0442 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.999292] env[62521]: DEBUG nova.compute.provider_tree [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.083920] env[62521]: DEBUG oslo_vmware.api [None req-f04c5b82-2026-45ab-a6fe-c4eb4f22343a tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318892, 'name': SuspendVM_Task} progress is 62%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.173783] env[62521]: DEBUG nova.network.neutron [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance_info_cache with network_info: [{"id": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "address": "fa:16:3e:8c:86:c8", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34cee820-50", "ovs_interfaceid": "34cee820-50a0-47bd-a0c7-25f1e01fa639", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.212516] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318891, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.661447} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.212909] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] b8158051-98b6-47d8-84c5-d1d56515afe8/b8158051-98b6-47d8-84c5-d1d56515afe8.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.213358] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.213708] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1a8c059f-687a-4381-a26b-12c39777bd4d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.220672] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 968.220672] env[62521]: value = "task-1318896" [ 968.220672] env[62521]: _type = "Task" [ 968.220672] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.231215] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318896, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.332169] env[62521]: DEBUG oslo_vmware.api [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318894, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.330083} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.332762] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.332965] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 968.333167] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 968.333351] env[62521]: INFO nova.compute.manager [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Took 1.16 seconds to destroy the instance on the hypervisor. [ 968.333597] env[62521]: DEBUG oslo.service.loopingcall [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.333795] env[62521]: DEBUG nova.compute.manager [-] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 968.333890] env[62521]: DEBUG nova.network.neutron [-] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 968.338250] env[62521]: DEBUG oslo_vmware.api [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318895, 'name': PowerOffVM_Task, 'duration_secs': 0.291276} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.338782] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 968.339052] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 968.339239] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dca59d9b-34b5-456a-bf4b-da57f6e1cc8f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.392044] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 968.392372] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7f2dc6c3-4f10-438c-bd53-67c422d5f1f1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.398257] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "refresh_cache-f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.398388] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "refresh_cache-f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.398530] env[62521]: DEBUG nova.network.neutron [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.402113] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 968.402113] env[62521]: value = "task-1318898" [ 968.402113] env[62521]: _type = "Task" [ 968.402113] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.411979] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318898, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.502557] env[62521]: DEBUG nova.scheduler.client.report [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.586150] env[62521]: DEBUG oslo_vmware.api [None req-f04c5b82-2026-45ab-a6fe-c4eb4f22343a tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318892, 'name': SuspendVM_Task, 'duration_secs': 0.961161} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.586450] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f04c5b82-2026-45ab-a6fe-c4eb4f22343a tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Suspended the VM {{(pid=62521) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 968.586700] env[62521]: DEBUG nova.compute.manager [None req-f04c5b82-2026-45ab-a6fe-c4eb4f22343a tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.587713] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5006a0a-2c7e-47a7-9f4d-9a6b97b5cb57 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.604024] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.604024] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.604024] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Deleting the datastore file [datastore1] d5f6492d-ca1b-4e74-b792-b04d55c33660 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.604024] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b440fe0-3489-4ced-bd87-3f12842a57a7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.610059] env[62521]: DEBUG oslo_vmware.api [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for the task: (returnval){ [ 968.610059] env[62521]: value = "task-1318899" [ 968.610059] env[62521]: _type = "Task" [ 968.610059] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.621172] env[62521]: DEBUG oslo_vmware.api [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318899, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.675643] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.675970] env[62521]: DEBUG nova.objects.instance [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lazy-loading 'migration_context' on Instance uuid 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.686029] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance_info_cache with network_info: [{"id": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "address": "fa:16:3e:d7:52:1f", "network": {"id": "b093f24c-ebce-4eb2-899e-c411d2fe54d3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c52a735a191b498eb13a1a29aec91182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc7aa55d-223a-4157-9137-88dc492f2db2", "external-id": "nsx-vlan-transportzone-290", "segmentation_id": 290, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e00f1c5-5c", "ovs_interfaceid": "7e00f1c5-5c15-430f-baff-cf464bf5d9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.731347] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318896, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.285349} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.731636] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.732440] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5639b83f-9e7c-4934-a495-3e0c6ab918f1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.755118] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] b8158051-98b6-47d8-84c5-d1d56515afe8/b8158051-98b6-47d8-84c5-d1d56515afe8.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.755417] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-289c912f-5144-4599-8f7c-e8f9099bd6b0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.776554] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 968.776554] env[62521]: value = "task-1318900" [ 968.776554] env[62521]: _type = "Task" [ 968.776554] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.784379] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318900, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.914408] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318898, 'name': CreateSnapshot_Task, 'duration_secs': 0.461076} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.914408] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 968.914408] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b9b4b9-e468-4f06-9383-b1467bda294b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.937810] env[62521]: DEBUG nova.network.neutron [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 969.010303] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.886s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.036969] env[62521]: INFO nova.scheduler.client.report [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleted allocations for instance af91de53-2ef8-48d1-b8c1-7bd0809f4ae4 [ 969.093760] env[62521]: DEBUG nova.network.neutron [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Updating instance_info_cache with network_info: [{"id": "45c9722b-0936-4e02-8f69-17a30523490b", "address": "fa:16:3e:07:01:02", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45c9722b-09", "ovs_interfaceid": "45c9722b-0936-4e02-8f69-17a30523490b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.095308] env[62521]: DEBUG nova.network.neutron [-] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.120939] env[62521]: DEBUG oslo_vmware.api [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Task: {'id': task-1318899, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.421533} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.121225] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.121412] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.121593] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.121767] env[62521]: INFO nova.compute.manager [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Took 1.33 seconds to destroy the instance on the hypervisor. [ 969.122017] env[62521]: DEBUG oslo.service.loopingcall [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.122454] env[62521]: DEBUG nova.compute.manager [-] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.122553] env[62521]: DEBUG nova.network.neutron [-] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 969.179122] env[62521]: DEBUG nova.objects.base [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Object Instance<02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c> lazy-loaded attributes: info_cache,migration_context {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 969.182103] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7df5aeb-ae31-41a5-9598-8f13a2c35d17 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.204766] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Releasing lock "refresh_cache-d5f6492d-ca1b-4e74-b792-b04d55c33660" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.204766] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updated the network info_cache for instance {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 969.205021] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.205487] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1912267-22e6-4cde-a4ec-d422f5d6fbaf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.208751] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.208855] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.209534] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.209702] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.209954] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.210026] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 969.210149] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.215680] env[62521]: DEBUG oslo_vmware.api [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 969.215680] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f13429-ff37-3910-7263-4775d0a22c43" [ 969.215680] env[62521]: _type = "Task" [ 969.215680] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.221929] env[62521]: DEBUG oslo_vmware.api [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f13429-ff37-3910-7263-4775d0a22c43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.287068] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318900, 'name': ReconfigVM_Task, 'duration_secs': 0.312924} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.287068] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Reconfigured VM instance instance-00000055 to attach disk [datastore2] b8158051-98b6-47d8-84c5-d1d56515afe8/b8158051-98b6-47d8-84c5-d1d56515afe8.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.287422] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-82f12daa-fd0d-4608-a138-7f0c1a7166c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.293549] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 969.293549] env[62521]: value = "task-1318901" [ 969.293549] env[62521]: _type = "Task" [ 969.293549] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.304216] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318901, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.433536] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 969.433878] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f49172ce-17b4-4c24-b885-22a8cbd4ad55 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.442452] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 969.442452] env[62521]: value = "task-1318902" [ 969.442452] env[62521]: _type = "Task" [ 969.442452] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.450765] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318902, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.469076] env[62521]: DEBUG nova.compute.manager [req-f6d789b3-4964-464c-be45-dcce221a508e req-e7f038e2-353c-4c8f-b9ea-4d4c1aa134c4 service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Received event network-vif-deleted-7e00f1c5-5c15-430f-baff-cf464bf5d9b3 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.469337] env[62521]: INFO nova.compute.manager [req-f6d789b3-4964-464c-be45-dcce221a508e req-e7f038e2-353c-4c8f-b9ea-4d4c1aa134c4 service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Neutron deleted interface 7e00f1c5-5c15-430f-baff-cf464bf5d9b3; detaching it from the instance and deleting it from the info cache [ 969.469463] env[62521]: DEBUG nova.network.neutron [req-f6d789b3-4964-464c-be45-dcce221a508e req-e7f038e2-353c-4c8f-b9ea-4d4c1aa134c4 service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.544319] env[62521]: DEBUG oslo_concurrency.lockutils [None req-dec4fb56-8634-4411-a6e8-647120ce9c42 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "af91de53-2ef8-48d1-b8c1-7bd0809f4ae4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.662s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.598085] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "refresh_cache-f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.598428] env[62521]: DEBUG nova.compute.manager [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Instance network_info: |[{"id": "45c9722b-0936-4e02-8f69-17a30523490b", "address": "fa:16:3e:07:01:02", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45c9722b-09", "ovs_interfaceid": "45c9722b-0936-4e02-8f69-17a30523490b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 969.598845] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:01:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '84aee122-f630-43c5-9cc1-3a38d3819c82', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '45c9722b-0936-4e02-8f69-17a30523490b', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.606433] env[62521]: DEBUG oslo.service.loopingcall [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.606851] env[62521]: INFO nova.compute.manager [-] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Took 1.27 seconds to deallocate network for instance. [ 969.607408] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 969.609308] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d71acb56-ae24-4400-ae95-0c021a988e97 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.634313] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.634313] env[62521]: value = "task-1318903" [ 969.634313] env[62521]: _type = "Task" [ 969.634313] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.643024] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318903, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.713333] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.713333] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.713537] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.713620] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 969.714738] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b712086-f6fc-4803-aafa-537c8bb08a8c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.730967] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f976aec-409e-4056-8829-e0d9b6079347 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.736050] env[62521]: DEBUG oslo_vmware.api [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f13429-ff37-3910-7263-4775d0a22c43, 'name': SearchDatastore_Task, 'duration_secs': 0.012921} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.736480] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.736842] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.749083] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77bb5b23-d424-4774-8a98-c65827044a2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.755191] env[62521]: DEBUG nova.compute.manager [req-a0c1acef-5423-4037-889a-257e0d752f5c req-c79f6ad1-c233-4e21-b460-021a86c9841f service nova] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Received event network-changed-45c9722b-0936-4e02-8f69-17a30523490b {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.755191] env[62521]: DEBUG nova.compute.manager [req-a0c1acef-5423-4037-889a-257e0d752f5c req-c79f6ad1-c233-4e21-b460-021a86c9841f service nova] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Refreshing instance network info cache due to event network-changed-45c9722b-0936-4e02-8f69-17a30523490b. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 969.755191] env[62521]: DEBUG oslo_concurrency.lockutils [req-a0c1acef-5423-4037-889a-257e0d752f5c req-c79f6ad1-c233-4e21-b460-021a86c9841f service nova] Acquiring lock "refresh_cache-f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.755379] env[62521]: DEBUG oslo_concurrency.lockutils [req-a0c1acef-5423-4037-889a-257e0d752f5c req-c79f6ad1-c233-4e21-b460-021a86c9841f service nova] Acquired lock "refresh_cache-f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.755447] env[62521]: DEBUG nova.network.neutron [req-a0c1acef-5423-4037-889a-257e0d752f5c req-c79f6ad1-c233-4e21-b460-021a86c9841f service nova] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Refreshing network info cache for port 45c9722b-0936-4e02-8f69-17a30523490b {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 969.759856] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266bdd85-e98c-433a-8fb3-38770c5ff365 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.789506] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179784MB free_disk=133GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 969.789657] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.804093] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318901, 'name': Rename_Task, 'duration_secs': 0.175351} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.804373] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.804621] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5182713b-a06b-4308-9c26-5662387e3cc4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.810470] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 969.810470] env[62521]: value = "task-1318904" [ 969.810470] env[62521]: _type = "Task" [ 969.810470] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.818384] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318904, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.884418] env[62521]: DEBUG nova.network.neutron [-] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.913944] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.914204] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.952139] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318902, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.972312] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f839bfd-aefd-4b08-98e4-d19621b671bb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.980913] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6861a367-da60-4a87-8bdc-a2a1b9b8f840 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.009149] env[62521]: DEBUG nova.compute.manager [req-f6d789b3-4964-464c-be45-dcce221a508e req-e7f038e2-353c-4c8f-b9ea-4d4c1aa134c4 service nova] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Detach interface failed, port_id=7e00f1c5-5c15-430f-baff-cf464bf5d9b3, reason: Instance d5f6492d-ca1b-4e74-b792-b04d55c33660 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 970.042034] env[62521]: INFO nova.compute.manager [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Resuming [ 970.042034] env[62521]: DEBUG nova.objects.instance [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lazy-loading 'flavor' on Instance uuid bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.132210] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.145693] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318903, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.326716] env[62521]: DEBUG oslo_vmware.api [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1318904, 'name': PowerOnVM_Task, 'duration_secs': 0.507234} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.327068] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.327320] env[62521]: INFO nova.compute.manager [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Took 6.92 seconds to spawn the instance on the hypervisor. [ 970.327599] env[62521]: DEBUG nova.compute.manager [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.328380] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f3bdc4-a047-48b1-b510-dcd5891ea73f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.387071] env[62521]: INFO nova.compute.manager [-] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Took 1.26 seconds to deallocate network for instance. [ 970.418155] env[62521]: DEBUG nova.compute.manager [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 970.447951] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93bcae8c-86c4-4dea-b323-2f581a35ca4d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.458947] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318902, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.464351] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c14cf6-b547-4b38-9b17-f073650b87b4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.499209] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3726cc05-ae66-43ff-bd0f-39710c62f182 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.507016] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc696604-de8a-4ccb-aac0-b3d0c9966334 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.520084] env[62521]: DEBUG nova.compute.provider_tree [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.536213] env[62521]: DEBUG nova.network.neutron [req-a0c1acef-5423-4037-889a-257e0d752f5c req-c79f6ad1-c233-4e21-b460-021a86c9841f service nova] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Updated VIF entry in instance network info cache for port 45c9722b-0936-4e02-8f69-17a30523490b. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.536611] env[62521]: DEBUG nova.network.neutron [req-a0c1acef-5423-4037-889a-257e0d752f5c req-c79f6ad1-c233-4e21-b460-021a86c9841f service nova] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Updating instance_info_cache with network_info: [{"id": "45c9722b-0936-4e02-8f69-17a30523490b", "address": "fa:16:3e:07:01:02", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45c9722b-09", "ovs_interfaceid": "45c9722b-0936-4e02-8f69-17a30523490b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.646411] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318903, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.849538] env[62521]: INFO nova.compute.manager [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Took 14.33 seconds to build instance. [ 970.896033] env[62521]: DEBUG oslo_concurrency.lockutils [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.938685] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.952881] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318902, 'name': CloneVM_Task} progress is 95%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.023828] env[62521]: DEBUG nova.scheduler.client.report [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.042482] env[62521]: DEBUG oslo_concurrency.lockutils [req-a0c1acef-5423-4037-889a-257e0d752f5c req-c79f6ad1-c233-4e21-b460-021a86c9841f service nova] Releasing lock "refresh_cache-f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.042669] env[62521]: DEBUG nova.compute.manager [req-a0c1acef-5423-4037-889a-257e0d752f5c req-c79f6ad1-c233-4e21-b460-021a86c9841f service nova] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Received event network-vif-deleted-fb88c8dc-e610-4e05-ba90-1a6dcf6ae085 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.049598] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.049764] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquired lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.050096] env[62521]: DEBUG nova.network.neutron [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 971.152728] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318903, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.351919] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3489a464-052e-4918-94f0-811b97a52231 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "b8158051-98b6-47d8-84c5-d1d56515afe8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.866s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.452455] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318902, 'name': CloneVM_Task, 'duration_secs': 1.783522} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.452774] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Created linked-clone VM from snapshot [ 971.453538] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8ef29e-ff78-42e7-a8ac-71f081be7d5e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.461014] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Uploading image 2572a161-29ce-4045-b2ab-d5198c3f86b7 {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 971.480958] env[62521]: DEBUG oslo_vmware.rw_handles [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 971.480958] env[62521]: value = "vm-282154" [ 971.480958] env[62521]: _type = "VirtualMachine" [ 971.480958] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 971.481439] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0bcaf032-a63c-4995-895c-514ab941b57b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.487362] env[62521]: DEBUG oslo_vmware.rw_handles [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lease: (returnval){ [ 971.487362] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5215fdcb-99f4-4b47-8bf5-9e983cc53588" [ 971.487362] env[62521]: _type = "HttpNfcLease" [ 971.487362] env[62521]: } obtained for exporting VM: (result){ [ 971.487362] env[62521]: value = "vm-282154" [ 971.487362] env[62521]: _type = "VirtualMachine" [ 971.487362] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 971.487753] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the lease: (returnval){ [ 971.487753] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5215fdcb-99f4-4b47-8bf5-9e983cc53588" [ 971.487753] env[62521]: _type = "HttpNfcLease" [ 971.487753] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 971.493645] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 971.493645] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5215fdcb-99f4-4b47-8bf5-9e983cc53588" [ 971.493645] env[62521]: _type = "HttpNfcLease" [ 971.493645] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 971.647918] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318903, 'name': CreateVM_Task, 'duration_secs': 1.804573} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.648190] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 971.648928] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.649133] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.649736] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 971.649893] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-648c2888-9b04-4407-8e1d-1404821c38ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.654753] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 971.654753] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5209fed4-9794-94c0-e814-cff2d7698652" [ 971.654753] env[62521]: _type = "Task" [ 971.654753] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.664023] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5209fed4-9794-94c0-e814-cff2d7698652, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.761679] env[62521]: DEBUG nova.network.neutron [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updating instance_info_cache with network_info: [{"id": "903b74d0-11d2-404d-81f1-b6bb18f713af", "address": "fa:16:3e:6a:82:e1", "network": {"id": "de64444b-3f14-4c6f-a085-6dae45fd818d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-159932846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a592d0f3c5834858bda6ca89a33662a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap903b74d0-11", "ovs_interfaceid": "903b74d0-11d2-404d-81f1-b6bb18f713af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.841414] env[62521]: DEBUG nova.compute.manager [req-3def5621-6d81-47d3-8ac4-6fcf2fb01a87 req-e2add48c-f15d-4f8b-a487-03a04c7ff363 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Received event network-changed-23b0cfbc-ec82-4a35-a248-249e86ecec84 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.841613] env[62521]: DEBUG nova.compute.manager [req-3def5621-6d81-47d3-8ac4-6fcf2fb01a87 req-e2add48c-f15d-4f8b-a487-03a04c7ff363 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Refreshing instance network info cache due to event network-changed-23b0cfbc-ec82-4a35-a248-249e86ecec84. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 971.841831] env[62521]: DEBUG oslo_concurrency.lockutils [req-3def5621-6d81-47d3-8ac4-6fcf2fb01a87 req-e2add48c-f15d-4f8b-a487-03a04c7ff363 service nova] Acquiring lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.841978] env[62521]: DEBUG oslo_concurrency.lockutils [req-3def5621-6d81-47d3-8ac4-6fcf2fb01a87 req-e2add48c-f15d-4f8b-a487-03a04c7ff363 service nova] Acquired lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.842156] env[62521]: DEBUG nova.network.neutron [req-3def5621-6d81-47d3-8ac4-6fcf2fb01a87 req-e2add48c-f15d-4f8b-a487-03a04c7ff363 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Refreshing network info cache for port 23b0cfbc-ec82-4a35-a248-249e86ecec84 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 971.995581] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 971.995581] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5215fdcb-99f4-4b47-8bf5-9e983cc53588" [ 971.995581] env[62521]: _type = "HttpNfcLease" [ 971.995581] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 972.078676] env[62521]: DEBUG oslo_vmware.rw_handles [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 972.078676] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5215fdcb-99f4-4b47-8bf5-9e983cc53588" [ 972.078676] env[62521]: _type = "HttpNfcLease" [ 972.078676] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 972.078676] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b22087-5432-46e2-890b-faad9565dc19 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.078676] env[62521]: DEBUG oslo_vmware.rw_handles [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52be166e-ebbc-ca8e-ea59-48c4048eb51a/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 972.078676] env[62521]: DEBUG oslo_vmware.rw_handles [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52be166e-ebbc-ca8e-ea59-48c4048eb51a/disk-0.vmdk for reading. {{(pid=62521) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 972.078676] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.326s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.078676] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.277s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.107290] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ddba12e5-0dcc-442a-8f48-ff7066add292 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.165436] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5209fed4-9794-94c0-e814-cff2d7698652, 'name': SearchDatastore_Task, 'duration_secs': 0.009488} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.165753] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.166070] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.166243] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.166391] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.167024] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.167024] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b96e0b77-82cf-4295-a566-b7c1c4427b34 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.174717] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.174914] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 972.175656] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce6b3172-f38b-4ab0-ad7a-b98ecac23c44 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.181035] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 972.181035] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52736cc3-28ee-ef71-6735-0e08cef49431" [ 972.181035] env[62521]: _type = "Task" [ 972.181035] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.188638] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52736cc3-28ee-ef71-6735-0e08cef49431, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.264347] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Releasing lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.264867] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397b48cb-3633-4e72-9565-cb8b5bea32f0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.274559] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Resuming the VM {{(pid=62521) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 972.278108] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-affbcc53-966b-4e61-b7a4-6b2d8d6aa675 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.282311] env[62521]: DEBUG oslo_vmware.api [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 972.282311] env[62521]: value = "task-1318906" [ 972.282311] env[62521]: _type = "Task" [ 972.282311] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.292807] env[62521]: DEBUG oslo_vmware.api [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318906, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.464040] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.465026] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.579034] env[62521]: DEBUG nova.network.neutron [req-3def5621-6d81-47d3-8ac4-6fcf2fb01a87 req-e2add48c-f15d-4f8b-a487-03a04c7ff363 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updated VIF entry in instance network info cache for port 23b0cfbc-ec82-4a35-a248-249e86ecec84. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 972.579467] env[62521]: DEBUG nova.network.neutron [req-3def5621-6d81-47d3-8ac4-6fcf2fb01a87 req-e2add48c-f15d-4f8b-a487-03a04c7ff363 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updating instance_info_cache with network_info: [{"id": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "address": "fa:16:3e:37:05:5a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b0cfbc-ec", "ovs_interfaceid": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.633647] env[62521]: INFO nova.scheduler.client.report [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted allocation for migration 371d8a8a-aab8-45be-aa37-32f199c69c75 [ 972.692122] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52736cc3-28ee-ef71-6735-0e08cef49431, 'name': SearchDatastore_Task, 'duration_secs': 0.010557} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.693271] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2743face-5b3d-4fa9-ba18-9f69cd323ca5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.698998] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 972.698998] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5256929e-8f3b-2dea-7e13-85985bd4391c" [ 972.698998] env[62521]: _type = "Task" [ 972.698998] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.708028] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5256929e-8f3b-2dea-7e13-85985bd4391c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.792713] env[62521]: DEBUG oslo_vmware.api [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318906, 'name': PowerOnVM_Task} progress is 96%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.967011] env[62521]: DEBUG nova.compute.manager [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 973.084172] env[62521]: DEBUG oslo_concurrency.lockutils [req-3def5621-6d81-47d3-8ac4-6fcf2fb01a87 req-e2add48c-f15d-4f8b-a487-03a04c7ff363 service nova] Releasing lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.107559] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.107559] env[62521]: WARNING nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance d5f6492d-ca1b-4e74-b792-b04d55c33660 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 973.107559] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance d574f077-b221-4f91-8b54-0915421cb36f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.109404] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 11e61e72-a311-4c43-bf53-df82ad9d70f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.109540] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.109704] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.110352] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.110548] env[62521]: WARNING nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 0743a5b6-87fe-4612-a5a1-5d6527b69eb2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 973.110652] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance b8158051-98b6-47d8-84c5-d1d56515afe8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.110783] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.140396] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d53e0233-05c5-4487-b6be-66a9d2e00538 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.849s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.210376] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5256929e-8f3b-2dea-7e13-85985bd4391c, 'name': SearchDatastore_Task, 'duration_secs': 0.011015} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.211137] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.211137] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d/f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.212017] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c472102-0b65-4135-a417-e6497a9c9252 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.218753] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 973.218753] env[62521]: value = "task-1318907" [ 973.218753] env[62521]: _type = "Task" [ 973.218753] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.226990] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318907, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.298331] env[62521]: DEBUG oslo_vmware.api [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318906, 'name': PowerOnVM_Task, 'duration_secs': 0.639158} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.298682] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Resumed the VM {{(pid=62521) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 973.299140] env[62521]: DEBUG nova.compute.manager [None req-5503dd8a-bffe-4b4d-8ccc-5d6090de234f tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.299997] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8053526f-3a0e-4dd6-aa63-a14a32959aa1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.496815] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.615639] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 973.729965] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318907, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.731583] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.731881] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.732121] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.732362] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.732674] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.739061] env[62521]: INFO nova.compute.manager [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Terminating instance [ 973.741768] env[62521]: DEBUG nova.compute.manager [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 973.742107] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 973.743264] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a14f5fb5-8157-4111-8a53-357684ac06e8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.752451] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.752751] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f1b7674e-addc-431a-891e-24e3eed3e900 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.759735] env[62521]: DEBUG oslo_vmware.api [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 973.759735] env[62521]: value = "task-1318908" [ 973.759735] env[62521]: _type = "Task" [ 973.759735] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.768913] env[62521]: DEBUG oslo_vmware.api [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318908, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.119334] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance a68d4cb7-bfad-48f6-8aeb-bca04248a0d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 974.119605] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 974.119765] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 974.230208] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318907, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5537} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.231203] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d/f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.231203] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.231387] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c6234693-c820-411f-90f4-6fd4aceb6158 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.239028] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 974.239028] env[62521]: value = "task-1318909" [ 974.239028] env[62521]: _type = "Task" [ 974.239028] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.247772] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.271807] env[62521]: DEBUG oslo_vmware.api [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318908, 'name': PowerOffVM_Task, 'duration_secs': 0.218189} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.272029] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.272203] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 974.272527] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fcebcf17-7e50-448c-a870-07bae1ae0a6b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.335253] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b690d78-9155-4dd9-a1d2-2c39b02f20a3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.342165] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 974.342464] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 974.342674] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleting the datastore file [datastore2] 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.345184] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec3ece1d-101b-4460-b339-40b700ae7545 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.347892] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3a8cac-99cb-4616-a5f2-afa4e9b81688 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.385422] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "a0646a00-1f1c-4892-bb32-6212c90e9e95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.385689] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "a0646a00-1f1c-4892-bb32-6212c90e9e95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.387246] env[62521]: DEBUG oslo_vmware.api [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 974.387246] env[62521]: value = "task-1318911" [ 974.387246] env[62521]: _type = "Task" [ 974.387246] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.388231] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978febd5-1e08-4a31-9996-a2b7ccb598d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.401438] env[62521]: DEBUG oslo_vmware.api [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318911, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.403381] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a402a09-73cf-440b-bcc6-e4c3e11ef4b0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.417985] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.749850] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.149263} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.750167] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.750983] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e52a8e-4412-46e8-a4a8-76b5bdb27d08 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.773924] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d/f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.774279] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2fa6a110-d0f1-40ba-b87c-6e8e0bb73e39 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.794679] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 974.794679] env[62521]: value = "task-1318912" [ 974.794679] env[62521]: _type = "Task" [ 974.794679] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.804276] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318912, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.889194] env[62521]: DEBUG nova.compute.manager [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 974.902174] env[62521]: DEBUG oslo_vmware.api [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318911, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.941591] env[62521]: ERROR nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [req-4e1b177d-b942-4832-8adf-8b01b39b9c8c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4e1b177d-b942-4832-8adf-8b01b39b9c8c"}]} [ 974.959315] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 974.975049] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 974.975259] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.989127] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 975.010968] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 975.304170] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8622d9d0-39d6-4801-9c0c-281b43cbebfd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.310224] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318912, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.315273] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6685ab08-5133-4afc-b102-04cd5a0b3486 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.357034] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8070bd5c-b9fc-46ac-af81-60c56ae6e8fc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.366644] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f503d4-f5fe-4dcb-8e4e-3b4a1defb550 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.380463] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 975.402503] env[62521]: DEBUG oslo_vmware.api [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318911, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.614658} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.404714] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 975.404910] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 975.405157] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 975.405276] env[62521]: INFO nova.compute.manager [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Took 1.66 seconds to destroy the instance on the hypervisor. [ 975.405521] env[62521]: DEBUG oslo.service.loopingcall [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.406156] env[62521]: DEBUG nova.compute.manager [-] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 975.406261] env[62521]: DEBUG nova.network.neutron [-] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 975.415904] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.708387] env[62521]: DEBUG nova.compute.manager [req-2819543d-d4f0-441f-9034-5f620ee1770f req-b85ebc30-71ce-4b21-91dc-12cdc72d809d service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Received event network-vif-deleted-34cee820-50a0-47bd-a0c7-25f1e01fa639 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.708539] env[62521]: INFO nova.compute.manager [req-2819543d-d4f0-441f-9034-5f620ee1770f req-b85ebc30-71ce-4b21-91dc-12cdc72d809d service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Neutron deleted interface 34cee820-50a0-47bd-a0c7-25f1e01fa639; detaching it from the instance and deleting it from the info cache [ 975.708737] env[62521]: DEBUG nova.network.neutron [req-2819543d-d4f0-441f-9034-5f620ee1770f req-b85ebc30-71ce-4b21-91dc-12cdc72d809d service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.805813] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318912, 'name': ReconfigVM_Task, 'duration_secs': 0.852599} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.806342] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Reconfigured VM instance instance-00000056 to attach disk [datastore1] f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d/f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.807035] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e60c780c-cd0a-4609-9e22-9b5c7b957e86 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.813853] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 975.813853] env[62521]: value = "task-1318913" [ 975.813853] env[62521]: _type = "Task" [ 975.813853] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.823473] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318913, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.917098] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 115 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 975.917380] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 115 to 116 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 975.917957] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 976.186064] env[62521]: DEBUG nova.network.neutron [-] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.211057] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e45fcdd1-65c2-487f-9fbb-66e258fedfe4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.221233] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4199dc-babf-42c5-936b-8c5de57c4fa6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.250613] env[62521]: DEBUG nova.compute.manager [req-2819543d-d4f0-441f-9034-5f620ee1770f req-b85ebc30-71ce-4b21-91dc-12cdc72d809d service nova] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Detach interface failed, port_id=34cee820-50a0-47bd-a0c7-25f1e01fa639, reason: Instance 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 976.324147] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318913, 'name': Rename_Task, 'duration_secs': 0.148182} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.324666] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 976.324890] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-842148e3-9aad-4943-a907-18726eb6e390 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.332367] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 976.332367] env[62521]: value = "task-1318914" [ 976.332367] env[62521]: _type = "Task" [ 976.332367] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.342780] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.422844] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 976.423088] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.356s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.423387] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.293s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.423582] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.425748] env[62521]: DEBUG oslo_concurrency.lockutils [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.530s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.425963] env[62521]: DEBUG oslo_concurrency.lockutils [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.427747] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.489s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.429336] env[62521]: INFO nova.compute.claims [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 976.451300] env[62521]: INFO nova.scheduler.client.report [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted allocations for instance 0743a5b6-87fe-4612-a5a1-5d6527b69eb2 [ 976.456336] env[62521]: INFO nova.scheduler.client.report [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Deleted allocations for instance d5f6492d-ca1b-4e74-b792-b04d55c33660 [ 976.689098] env[62521]: INFO nova.compute.manager [-] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Took 1.28 seconds to deallocate network for instance. [ 976.844565] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318914, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.964598] env[62521]: DEBUG oslo_concurrency.lockutils [None req-687818ec-8513-4c46-9441-63aa9e6fca83 tempest-MigrationsAdminTest-323195408 tempest-MigrationsAdminTest-323195408-project-member] Lock "d5f6492d-ca1b-4e74-b792-b04d55c33660" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.180s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.966367] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b9471384-937c-4bf4-aa94-f6d17ed9b10d tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "0743a5b6-87fe-4612-a5a1-5d6527b69eb2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.806s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.195857] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.345395] env[62521]: DEBUG oslo_vmware.api [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318914, 'name': PowerOnVM_Task, 'duration_secs': 0.587117} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.345930] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 977.345930] env[62521]: INFO nova.compute.manager [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Took 10.60 seconds to spawn the instance on the hypervisor. [ 977.346107] env[62521]: DEBUG nova.compute.manager [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.347698] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ebcb6e-cfb8-4312-8bb7-44c83cdae0b3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.613451] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9363ce9-b59f-4848-b7b2-093574a821ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.621654] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf7c13f-604e-4913-a277-ad59f602f112 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.655754] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9676633d-3e26-4810-88fc-8302b986cda0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.663736] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28eef255-42f0-48b9-bbf1-903e82094342 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.677971] env[62521]: DEBUG nova.compute.provider_tree [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.780527] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.780898] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.868596] env[62521]: INFO nova.compute.manager [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Took 16.43 seconds to build instance. [ 978.181660] env[62521]: DEBUG nova.scheduler.client.report [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.283646] env[62521]: DEBUG nova.compute.manager [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 978.370710] env[62521]: DEBUG oslo_concurrency.lockutils [None req-551d08a5-3a8d-4aeb-987e-9ef1f010b987 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.939s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.687579] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.688174] env[62521]: DEBUG nova.compute.manager [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.691605] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.195s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.693318] env[62521]: INFO nova.compute.claims [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.804974] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.066845] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.067225] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.067396] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.067587] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.067788] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.070130] env[62521]: INFO nova.compute.manager [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Terminating instance [ 979.072074] env[62521]: DEBUG nova.compute.manager [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 979.072275] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 979.073200] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb21ded7-9b73-4b9b-92a5-d9f986a21a35 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.081801] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.082494] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f52ba646-4187-40f7-a60c-ca11d4913bf8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.089024] env[62521]: DEBUG oslo_vmware.api [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 979.089024] env[62521]: value = "task-1318915" [ 979.089024] env[62521]: _type = "Task" [ 979.089024] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.097857] env[62521]: DEBUG oslo_vmware.api [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318915, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.193842] env[62521]: DEBUG nova.compute.utils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.195380] env[62521]: DEBUG nova.compute.manager [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 979.195609] env[62521]: DEBUG nova.network.neutron [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 979.247073] env[62521]: DEBUG nova.policy [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a28a24310a04c6d861292815c825dac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f029ddf5cd8448caad5ec23775b6503', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 979.277322] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "69c5ff30-259d-4067-b5e2-ab727009d8ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.277681] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "69c5ff30-259d-4067-b5e2-ab727009d8ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.526502] env[62521]: DEBUG nova.network.neutron [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Successfully created port: dee03f5c-2442-438b-8786-6f3e0c35f8ab {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.602515] env[62521]: DEBUG oslo_vmware.api [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318915, 'name': PowerOffVM_Task, 'duration_secs': 0.210946} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.602852] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.603245] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 979.603610] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a6c4a05-dc24-481c-9f28-18bcca572cbe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.707154] env[62521]: DEBUG nova.compute.manager [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 979.784283] env[62521]: DEBUG nova.compute.manager [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 979.942289] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b30819a-0ba2-45c8-a330-dae267f34d95 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.950522] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22f4d68-d3fd-4a9b-8d7a-62a68c38c474 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.981667] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0c715d-cbac-41d8-a066-79f24ada5dcd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.989578] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d99476-55f1-4a39-b4df-c495d04598b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.003478] env[62521]: DEBUG nova.compute.provider_tree [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.109496] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 980.110077] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 980.110077] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleting the datastore file [datastore1] bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.110306] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f02ee71-9016-4818-b004-1debf05d1f02 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.117487] env[62521]: DEBUG oslo_vmware.api [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for the task: (returnval){ [ 980.117487] env[62521]: value = "task-1318917" [ 980.117487] env[62521]: _type = "Task" [ 980.117487] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.126485] env[62521]: DEBUG oslo_vmware.api [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.313294] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.453361] env[62521]: DEBUG oslo_vmware.rw_handles [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52be166e-ebbc-ca8e-ea59-48c4048eb51a/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 980.455945] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b9e3a7-b746-4dcd-9948-299030f6be49 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.462046] env[62521]: DEBUG oslo_vmware.rw_handles [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52be166e-ebbc-ca8e-ea59-48c4048eb51a/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 980.462233] env[62521]: ERROR oslo_vmware.rw_handles [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52be166e-ebbc-ca8e-ea59-48c4048eb51a/disk-0.vmdk due to incomplete transfer. [ 980.462481] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e121ffbd-4134-41b6-86b6-ccbd9cc77443 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.469943] env[62521]: DEBUG oslo_vmware.rw_handles [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52be166e-ebbc-ca8e-ea59-48c4048eb51a/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 980.470042] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Uploaded image 2572a161-29ce-4045-b2ab-d5198c3f86b7 to the Glance image server {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 980.472475] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 980.472728] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-762ca707-c57b-46ce-a603-87ff70d081b8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.479109] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 980.479109] env[62521]: value = "task-1318918" [ 980.479109] env[62521]: _type = "Task" [ 980.479109] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.489042] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318918, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.506268] env[62521]: DEBUG nova.scheduler.client.report [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.633111] env[62521]: DEBUG oslo_vmware.api [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.718608] env[62521]: DEBUG nova.compute.manager [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 980.742053] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.742053] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.742053] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.742053] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.742053] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.742627] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.743038] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.743365] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.743714] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.744044] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.744397] env[62521]: DEBUG nova.virt.hardware [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.745550] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-938c4212-7ea1-4f72-be09-2364f5854704 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.757251] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efa265a-22e2-4f25-bf32-9cc316228037 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.962620] env[62521]: DEBUG nova.compute.manager [req-472a474a-ae2f-47ae-9ad9-0849d5fef63f req-ab389b02-5c88-4c6c-b08e-c32864266ce5 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Received event network-vif-plugged-dee03f5c-2442-438b-8786-6f3e0c35f8ab {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.962868] env[62521]: DEBUG oslo_concurrency.lockutils [req-472a474a-ae2f-47ae-9ad9-0849d5fef63f req-ab389b02-5c88-4c6c-b08e-c32864266ce5 service nova] Acquiring lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.963098] env[62521]: DEBUG oslo_concurrency.lockutils [req-472a474a-ae2f-47ae-9ad9-0849d5fef63f req-ab389b02-5c88-4c6c-b08e-c32864266ce5 service nova] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.963293] env[62521]: DEBUG oslo_concurrency.lockutils [req-472a474a-ae2f-47ae-9ad9-0849d5fef63f req-ab389b02-5c88-4c6c-b08e-c32864266ce5 service nova] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.963450] env[62521]: DEBUG nova.compute.manager [req-472a474a-ae2f-47ae-9ad9-0849d5fef63f req-ab389b02-5c88-4c6c-b08e-c32864266ce5 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] No waiting events found dispatching network-vif-plugged-dee03f5c-2442-438b-8786-6f3e0c35f8ab {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 980.963619] env[62521]: WARNING nova.compute.manager [req-472a474a-ae2f-47ae-9ad9-0849d5fef63f req-ab389b02-5c88-4c6c-b08e-c32864266ce5 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Received unexpected event network-vif-plugged-dee03f5c-2442-438b-8786-6f3e0c35f8ab for instance with vm_state building and task_state spawning. [ 980.988528] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318918, 'name': Destroy_Task, 'duration_secs': 0.332546} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.988695] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Destroyed the VM [ 980.988954] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 980.989986] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a8f4d5c4-7fe3-49d5-a977-ddbba097bc55 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.995398] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 980.995398] env[62521]: value = "task-1318919" [ 980.995398] env[62521]: _type = "Task" [ 980.995398] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.004013] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318919, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.010997] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.011589] env[62521]: DEBUG nova.compute.manager [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 981.015201] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.599s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.016793] env[62521]: INFO nova.compute.claims [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.067893] env[62521]: DEBUG nova.network.neutron [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Successfully updated port: dee03f5c-2442-438b-8786-6f3e0c35f8ab {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 981.129183] env[62521]: DEBUG oslo_vmware.api [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Task: {'id': task-1318917, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.552893} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.129558] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.129877] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 981.130136] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.130331] env[62521]: INFO nova.compute.manager [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Took 2.06 seconds to destroy the instance on the hypervisor. [ 981.130584] env[62521]: DEBUG oslo.service.loopingcall [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.130805] env[62521]: DEBUG nova.compute.manager [-] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 981.130906] env[62521]: DEBUG nova.network.neutron [-] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 981.411615] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquiring lock "31f926a8-777d-4882-afd6-1e4dba3f4e11" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.411998] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lock "31f926a8-777d-4882-afd6-1e4dba3f4e11" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.505495] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318919, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.521221] env[62521]: DEBUG nova.compute.utils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.524549] env[62521]: DEBUG nova.compute.manager [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 981.524738] env[62521]: DEBUG nova.network.neutron [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 981.564965] env[62521]: DEBUG nova.policy [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba9b0b82d53e4502bef18a9f031fcc9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5780fc8ecfd64161bc2ea0b3a092b4d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 981.570664] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "refresh_cache-7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.570802] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquired lock "refresh_cache-7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.570948] env[62521]: DEBUG nova.network.neutron [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.834339] env[62521]: DEBUG nova.network.neutron [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Successfully created port: add56f48-9368-4591-b8b6-cc7be1121d28 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.911550] env[62521]: DEBUG nova.network.neutron [-] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.916141] env[62521]: DEBUG nova.compute.manager [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 982.005335] env[62521]: DEBUG oslo_vmware.api [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318919, 'name': RemoveSnapshot_Task, 'duration_secs': 0.644758} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.005632] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 982.005862] env[62521]: INFO nova.compute.manager [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Took 14.16 seconds to snapshot the instance on the hypervisor. [ 982.025846] env[62521]: DEBUG nova.compute.manager [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 982.122787] env[62521]: DEBUG nova.network.neutron [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 982.255239] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691c45a9-ce4c-4412-b4b9-7b2eacdbff24 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.263050] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3beb0d-9ec6-473e-99b1-eff7e801d2d1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.300024] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3daf84-4a01-4cb9-b5f2-37bdc7a34610 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.306853] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637ef874-771e-41d1-b199-6c37a69e443a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.320343] env[62521]: DEBUG nova.compute.provider_tree [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.371921] env[62521]: DEBUG nova.network.neutron [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Updating instance_info_cache with network_info: [{"id": "dee03f5c-2442-438b-8786-6f3e0c35f8ab", "address": "fa:16:3e:64:68:af", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdee03f5c-24", "ovs_interfaceid": "dee03f5c-2442-438b-8786-6f3e0c35f8ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.414538] env[62521]: INFO nova.compute.manager [-] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Took 1.28 seconds to deallocate network for instance. [ 982.437467] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.549442] env[62521]: DEBUG nova.compute.manager [None req-53c966a8-0b6e-4142-b42b-b77cbcec6ea6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Found 2 images (rotation: 2) {{(pid=62521) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 982.823511] env[62521]: DEBUG nova.scheduler.client.report [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.875486] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Releasing lock "refresh_cache-7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.875810] env[62521]: DEBUG nova.compute.manager [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Instance network_info: |[{"id": "dee03f5c-2442-438b-8786-6f3e0c35f8ab", "address": "fa:16:3e:64:68:af", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdee03f5c-24", "ovs_interfaceid": "dee03f5c-2442-438b-8786-6f3e0c35f8ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.876243] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:68:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dee03f5c-2442-438b-8786-6f3e0c35f8ab', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.883691] env[62521]: DEBUG oslo.service.loopingcall [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.883902] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.884164] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82aa5250-955f-40c8-b305-4af62d46f291 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.905540] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.905540] env[62521]: value = "task-1318920" [ 982.905540] env[62521]: _type = "Task" [ 982.905540] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.913980] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318920, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.925218] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.993858] env[62521]: DEBUG nova.compute.manager [req-2aace0ae-9e47-467c-a6fc-6c3cf22a71ec req-61f7dd15-b7e7-4c93-996e-db4d134293e1 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Received event network-changed-dee03f5c-2442-438b-8786-6f3e0c35f8ab {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.994095] env[62521]: DEBUG nova.compute.manager [req-2aace0ae-9e47-467c-a6fc-6c3cf22a71ec req-61f7dd15-b7e7-4c93-996e-db4d134293e1 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Refreshing instance network info cache due to event network-changed-dee03f5c-2442-438b-8786-6f3e0c35f8ab. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.994363] env[62521]: DEBUG oslo_concurrency.lockutils [req-2aace0ae-9e47-467c-a6fc-6c3cf22a71ec req-61f7dd15-b7e7-4c93-996e-db4d134293e1 service nova] Acquiring lock "refresh_cache-7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.994458] env[62521]: DEBUG oslo_concurrency.lockutils [req-2aace0ae-9e47-467c-a6fc-6c3cf22a71ec req-61f7dd15-b7e7-4c93-996e-db4d134293e1 service nova] Acquired lock "refresh_cache-7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.994691] env[62521]: DEBUG nova.network.neutron [req-2aace0ae-9e47-467c-a6fc-6c3cf22a71ec req-61f7dd15-b7e7-4c93-996e-db4d134293e1 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Refreshing network info cache for port dee03f5c-2442-438b-8786-6f3e0c35f8ab {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 983.042941] env[62521]: DEBUG nova.compute.manager [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.064144] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.064414] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.064579] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.064786] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.064948] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.065199] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.065422] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.065617] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.065799] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.065970] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.066180] env[62521]: DEBUG nova.virt.hardware [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.067068] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0d89dd-b09d-4f7c-bdd2-18114ca646fb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.075341] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc54b2c-cea2-4d9a-9c01-69d2a78771de {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.329737] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.330323] env[62521]: DEBUG nova.compute.manager [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.332869] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.137s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.336020] env[62521]: DEBUG nova.objects.instance [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lazy-loading 'resources' on Instance uuid 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.351070] env[62521]: DEBUG nova.network.neutron [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Successfully updated port: add56f48-9368-4591-b8b6-cc7be1121d28 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.417035] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318920, 'name': CreateVM_Task, 'duration_secs': 0.338226} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.417225] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 983.418044] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.418153] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.418486] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.418776] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b6ce166-f66c-4380-ad53-b7301777a4d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.423695] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 983.423695] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52452d01-df58-a873-46fa-58cc878c3f12" [ 983.423695] env[62521]: _type = "Task" [ 983.423695] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.431209] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52452d01-df58-a873-46fa-58cc878c3f12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.505415] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "aa266be6-2bf1-445f-a968-2c144d30ecfa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.505662] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "aa266be6-2bf1-445f-a968-2c144d30ecfa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.687425] env[62521]: DEBUG nova.network.neutron [req-2aace0ae-9e47-467c-a6fc-6c3cf22a71ec req-61f7dd15-b7e7-4c93-996e-db4d134293e1 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Updated VIF entry in instance network info cache for port dee03f5c-2442-438b-8786-6f3e0c35f8ab. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.687805] env[62521]: DEBUG nova.network.neutron [req-2aace0ae-9e47-467c-a6fc-6c3cf22a71ec req-61f7dd15-b7e7-4c93-996e-db4d134293e1 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Updating instance_info_cache with network_info: [{"id": "dee03f5c-2442-438b-8786-6f3e0c35f8ab", "address": "fa:16:3e:64:68:af", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdee03f5c-24", "ovs_interfaceid": "dee03f5c-2442-438b-8786-6f3e0c35f8ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.835367] env[62521]: DEBUG nova.compute.utils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.839649] env[62521]: DEBUG nova.compute.manager [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 983.839832] env[62521]: DEBUG nova.network.neutron [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 983.854277] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "refresh_cache-a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.854277] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "refresh_cache-a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.854277] env[62521]: DEBUG nova.network.neutron [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 983.880644] env[62521]: DEBUG nova.policy [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d7aab6f3f6245c2b56979677428a9df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e6baab1556e4b00b07ab554664b4445', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 983.934633] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52452d01-df58-a873-46fa-58cc878c3f12, 'name': SearchDatastore_Task, 'duration_secs': 0.009176} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.937296] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.937558] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.937898] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.937975] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.938131] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.939275] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95ac3774-3c0a-4e1e-b6c7-e18eb9afe9c4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.947408] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.947601] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.948359] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87e3ed06-dfb8-4be8-8fe3-7e93efe20799 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.956045] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 983.956045] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52dad292-da16-78a5-1856-b418e7d6d976" [ 983.956045] env[62521]: _type = "Task" [ 983.956045] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.964174] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52dad292-da16-78a5-1856-b418e7d6d976, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.978033] env[62521]: DEBUG nova.compute.manager [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.978947] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a5730f-5c27-4b5c-a153-1378142a60ab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.007788] env[62521]: DEBUG nova.compute.manager [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 984.070709] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544be800-1442-4a2e-b3f2-e673a8a507d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.078775] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef360cb2-6a66-4e0d-bd09-256d073f6f0c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.112691] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231df920-409e-4abb-a1a8-44f7b172c463 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.123273] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4ff457-f26a-4fba-8d66-1ea272ae133c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.139545] env[62521]: DEBUG nova.compute.provider_tree [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.149933] env[62521]: DEBUG nova.network.neutron [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Successfully created port: ff67aa5f-1a3c-478f-bf15-cf0577756a05 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.190159] env[62521]: DEBUG oslo_concurrency.lockutils [req-2aace0ae-9e47-467c-a6fc-6c3cf22a71ec req-61f7dd15-b7e7-4c93-996e-db4d134293e1 service nova] Releasing lock "refresh_cache-7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.190461] env[62521]: DEBUG nova.compute.manager [req-2aace0ae-9e47-467c-a6fc-6c3cf22a71ec req-61f7dd15-b7e7-4c93-996e-db4d134293e1 service nova] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Received event network-vif-deleted-903b74d0-11d2-404d-81f1-b6bb18f713af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.342449] env[62521]: DEBUG nova.compute.manager [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.386586] env[62521]: DEBUG nova.network.neutron [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 984.466695] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52dad292-da16-78a5-1856-b418e7d6d976, 'name': SearchDatastore_Task, 'duration_secs': 0.00973} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.467516] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dee6b7d1-da04-4f60-b11d-eacd458d82ec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.472239] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 984.472239] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bc7c12-7810-81ce-3f55-32f6322dc16c" [ 984.472239] env[62521]: _type = "Task" [ 984.472239] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.479347] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bc7c12-7810-81ce-3f55-32f6322dc16c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.490316] env[62521]: INFO nova.compute.manager [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] instance snapshotting [ 984.490883] env[62521]: DEBUG nova.objects.instance [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'flavor' on Instance uuid 11e61e72-a311-4c43-bf53-df82ad9d70f7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.512080] env[62521]: DEBUG nova.network.neutron [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Updating instance_info_cache with network_info: [{"id": "add56f48-9368-4591-b8b6-cc7be1121d28", "address": "fa:16:3e:f2:8e:fd", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadd56f48-93", "ovs_interfaceid": "add56f48-9368-4591-b8b6-cc7be1121d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.531957] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.642308] env[62521]: DEBUG nova.scheduler.client.report [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.985032] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bc7c12-7810-81ce-3f55-32f6322dc16c, 'name': SearchDatastore_Task, 'duration_secs': 0.009479} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.985032] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.985032] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c/7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 984.985032] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e3ecd42-b30e-4b9f-80d5-689d9750e73f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.993974] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 984.993974] env[62521]: value = "task-1318921" [ 984.993974] env[62521]: _type = "Task" [ 984.993974] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.000560] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88edb791-1371-4244-8ded-23d2c9e832b8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.006828] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318921, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.026137] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "refresh_cache-a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.026137] env[62521]: DEBUG nova.compute.manager [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Instance network_info: |[{"id": "add56f48-9368-4591-b8b6-cc7be1121d28", "address": "fa:16:3e:f2:8e:fd", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadd56f48-93", "ovs_interfaceid": "add56f48-9368-4591-b8b6-cc7be1121d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 985.026137] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:8e:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd998416-f3d6-4a62-b828-5011063ce76a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'add56f48-9368-4591-b8b6-cc7be1121d28', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.032667] env[62521]: DEBUG oslo.service.loopingcall [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.033419] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4d3879-e95e-4627-a4b6-036ae3aba2e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.037230] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.037470] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3846b167-e028-46a7-be2e-eb3484f3ffab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.060258] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.060258] env[62521]: value = "task-1318922" [ 985.060258] env[62521]: _type = "Task" [ 985.060258] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.067046] env[62521]: DEBUG nova.compute.manager [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Received event network-vif-plugged-add56f48-9368-4591-b8b6-cc7be1121d28 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.067509] env[62521]: DEBUG oslo_concurrency.lockutils [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] Acquiring lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.067735] env[62521]: DEBUG oslo_concurrency.lockutils [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.067919] env[62521]: DEBUG oslo_concurrency.lockutils [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.068117] env[62521]: DEBUG nova.compute.manager [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] No waiting events found dispatching network-vif-plugged-add56f48-9368-4591-b8b6-cc7be1121d28 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 985.068299] env[62521]: WARNING nova.compute.manager [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Received unexpected event network-vif-plugged-add56f48-9368-4591-b8b6-cc7be1121d28 for instance with vm_state building and task_state spawning. [ 985.068479] env[62521]: DEBUG nova.compute.manager [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Received event network-changed-add56f48-9368-4591-b8b6-cc7be1121d28 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.068629] env[62521]: DEBUG nova.compute.manager [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Refreshing instance network info cache due to event network-changed-add56f48-9368-4591-b8b6-cc7be1121d28. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 985.068819] env[62521]: DEBUG oslo_concurrency.lockutils [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] Acquiring lock "refresh_cache-a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.068962] env[62521]: DEBUG oslo_concurrency.lockutils [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] Acquired lock "refresh_cache-a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.069173] env[62521]: DEBUG nova.network.neutron [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Refreshing network info cache for port add56f48-9368-4591-b8b6-cc7be1121d28 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.077082] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318922, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.148070] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.815s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.150041] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.345s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.151587] env[62521]: INFO nova.compute.claims [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.170582] env[62521]: INFO nova.scheduler.client.report [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted allocations for instance 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c [ 985.352433] env[62521]: DEBUG nova.compute.manager [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.381601] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.381869] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.382049] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.382247] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.382397] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.382550] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.382764] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.382928] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.383118] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.383294] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.383470] env[62521]: DEBUG nova.virt.hardware [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.384392] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd54b93b-0dcb-492f-a5f7-6defc034c576 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.393278] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b030f06b-795f-46b6-bbf4-8c94a548ec5a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.501479] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318921, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.554568] env[62521]: DEBUG nova.compute.manager [req-b83fe186-36db-4442-b06c-cd4b36d1d5d8 req-deee3712-eb76-43a3-a297-cdde2399667d service nova] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Received event network-vif-plugged-ff67aa5f-1a3c-478f-bf15-cf0577756a05 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.554847] env[62521]: DEBUG oslo_concurrency.lockutils [req-b83fe186-36db-4442-b06c-cd4b36d1d5d8 req-deee3712-eb76-43a3-a297-cdde2399667d service nova] Acquiring lock "a0646a00-1f1c-4892-bb32-6212c90e9e95-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.555108] env[62521]: DEBUG oslo_concurrency.lockutils [req-b83fe186-36db-4442-b06c-cd4b36d1d5d8 req-deee3712-eb76-43a3-a297-cdde2399667d service nova] Lock "a0646a00-1f1c-4892-bb32-6212c90e9e95-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.555287] env[62521]: DEBUG oslo_concurrency.lockutils [req-b83fe186-36db-4442-b06c-cd4b36d1d5d8 req-deee3712-eb76-43a3-a297-cdde2399667d service nova] Lock "a0646a00-1f1c-4892-bb32-6212c90e9e95-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.555464] env[62521]: DEBUG nova.compute.manager [req-b83fe186-36db-4442-b06c-cd4b36d1d5d8 req-deee3712-eb76-43a3-a297-cdde2399667d service nova] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] No waiting events found dispatching network-vif-plugged-ff67aa5f-1a3c-478f-bf15-cf0577756a05 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 985.555636] env[62521]: WARNING nova.compute.manager [req-b83fe186-36db-4442-b06c-cd4b36d1d5d8 req-deee3712-eb76-43a3-a297-cdde2399667d service nova] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Received unexpected event network-vif-plugged-ff67aa5f-1a3c-478f-bf15-cf0577756a05 for instance with vm_state building and task_state spawning. [ 985.562511] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 985.563434] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-75a42b0a-7ff3-4100-8e7e-a6620b9bf30f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.574660] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318922, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.576231] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 985.576231] env[62521]: value = "task-1318923" [ 985.576231] env[62521]: _type = "Task" [ 985.576231] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.593710] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318923, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.682754] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab14472e-d3c0-49be-8d77-d399f597d35c tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.951s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.836465] env[62521]: DEBUG nova.network.neutron [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Updated VIF entry in instance network info cache for port add56f48-9368-4591-b8b6-cc7be1121d28. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 985.836874] env[62521]: DEBUG nova.network.neutron [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Updating instance_info_cache with network_info: [{"id": "add56f48-9368-4591-b8b6-cc7be1121d28", "address": "fa:16:3e:f2:8e:fd", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadd56f48-93", "ovs_interfaceid": "add56f48-9368-4591-b8b6-cc7be1121d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.002848] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318921, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511295} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.003141] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c/7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 986.003365] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.003626] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd674906-1802-42d9-9664-d49a70024e45 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.010028] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 986.010028] env[62521]: value = "task-1318924" [ 986.010028] env[62521]: _type = "Task" [ 986.010028] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.018463] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318924, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.073029] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318922, 'name': CreateVM_Task, 'duration_secs': 0.591883} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.073262] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 986.073910] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.074096] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.074442] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.075077] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7702369-9326-4802-abde-6386f5b255b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.083536] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 986.083536] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521466be-a73d-ba32-64a2-8b84bcbfa555" [ 986.083536] env[62521]: _type = "Task" [ 986.083536] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.086550] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318923, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.094116] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521466be-a73d-ba32-64a2-8b84bcbfa555, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.254029] env[62521]: DEBUG nova.network.neutron [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Successfully updated port: ff67aa5f-1a3c-478f-bf15-cf0577756a05 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.340313] env[62521]: DEBUG oslo_concurrency.lockutils [req-9b756bfd-3014-4b8e-879b-d17a14aa79b5 req-fe4d539e-342f-48af-81f5-68351cef423d service nova] Releasing lock "refresh_cache-a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.375266] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413fd755-2c28-438a-9aa9-d163487793b5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.383257] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953827a3-be45-409f-83ce-343518095c80 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.413947] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906cefc6-236c-42ba-8dc3-64224ffbb357 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.421301] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28eec9ab-6e61-4ae5-a660-93f398eea901 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.435634] env[62521]: DEBUG nova.compute.provider_tree [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.519098] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318924, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068681} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.519391] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 986.520134] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311ac206-a035-4fbf-b33d-7c8a1d8fd296 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.541049] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c/7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.541307] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90a0c90b-2edd-4e7c-8b70-2a3829ae904f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.560323] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 986.560323] env[62521]: value = "task-1318925" [ 986.560323] env[62521]: _type = "Task" [ 986.560323] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.567654] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318925, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.586286] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318923, 'name': CreateSnapshot_Task, 'duration_secs': 0.74969} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.586626] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 986.590261] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae73c75f-ab61-4626-a9b7-33e56968e308 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.598210] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521466be-a73d-ba32-64a2-8b84bcbfa555, 'name': SearchDatastore_Task, 'duration_secs': 0.017811} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.601807] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.602066] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.602309] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.602461] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.602643] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.605013] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34f2c35b-d820-4212-9e86-600948f2a9ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.615880] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.616084] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 986.617032] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed6d4705-7a22-4de6-9c60-6d3862e013dd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.622179] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 986.622179] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52311158-a63d-7654-7e3a-b88350c3f3b6" [ 986.622179] env[62521]: _type = "Task" [ 986.622179] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.629025] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52311158-a63d-7654-7e3a-b88350c3f3b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.757023] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "refresh_cache-a0646a00-1f1c-4892-bb32-6212c90e9e95" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.757194] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "refresh_cache-a0646a00-1f1c-4892-bb32-6212c90e9e95" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.757352] env[62521]: DEBUG nova.network.neutron [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.939072] env[62521]: DEBUG nova.scheduler.client.report [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.071144] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318925, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.091293] env[62521]: DEBUG nova.compute.manager [req-1dab3caa-a5ce-42cc-894b-f06cabe77cdb req-79a9235e-e36f-446e-b9ad-f168ccfcdd8b service nova] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Received event network-changed-ff67aa5f-1a3c-478f-bf15-cf0577756a05 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.091521] env[62521]: DEBUG nova.compute.manager [req-1dab3caa-a5ce-42cc-894b-f06cabe77cdb req-79a9235e-e36f-446e-b9ad-f168ccfcdd8b service nova] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Refreshing instance network info cache due to event network-changed-ff67aa5f-1a3c-478f-bf15-cf0577756a05. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 987.091751] env[62521]: DEBUG oslo_concurrency.lockutils [req-1dab3caa-a5ce-42cc-894b-f06cabe77cdb req-79a9235e-e36f-446e-b9ad-f168ccfcdd8b service nova] Acquiring lock "refresh_cache-a0646a00-1f1c-4892-bb32-6212c90e9e95" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.113890] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 987.114418] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cfd00ea1-4569-4017-a9d0-4ce308020162 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.123052] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 987.123052] env[62521]: value = "task-1318926" [ 987.123052] env[62521]: _type = "Task" [ 987.123052] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.135422] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52311158-a63d-7654-7e3a-b88350c3f3b6, 'name': SearchDatastore_Task, 'duration_secs': 0.039052} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.139051] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318926, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.139273] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4db1372-8826-41ba-888c-8ec48271a082 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.144341] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 987.144341] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52430c0c-1399-1024-d146-927a7f3204f4" [ 987.144341] env[62521]: _type = "Task" [ 987.144341] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.151664] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52430c0c-1399-1024-d146-927a7f3204f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.290272] env[62521]: DEBUG nova.network.neutron [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 987.438765] env[62521]: DEBUG nova.network.neutron [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Updating instance_info_cache with network_info: [{"id": "ff67aa5f-1a3c-478f-bf15-cf0577756a05", "address": "fa:16:3e:65:50:04", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff67aa5f-1a", "ovs_interfaceid": "ff67aa5f-1a3c-478f-bf15-cf0577756a05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.443634] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.294s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.444373] env[62521]: DEBUG nova.compute.manager [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 987.447048] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.134s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.448569] env[62521]: INFO nova.compute.claims [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.570641] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318925, 'name': ReconfigVM_Task, 'duration_secs': 0.62844} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.570899] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c/7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.571587] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-795aaaa0-ec84-4ede-bd82-87355aea66d4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.578520] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 987.578520] env[62521]: value = "task-1318927" [ 987.578520] env[62521]: _type = "Task" [ 987.578520] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.586187] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318927, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.637929] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318926, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.654067] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52430c0c-1399-1024-d146-927a7f3204f4, 'name': SearchDatastore_Task, 'duration_secs': 0.010787} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.654067] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.654489] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] a68d4cb7-bfad-48f6-8aeb-bca04248a0d7/a68d4cb7-bfad-48f6-8aeb-bca04248a0d7.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.654620] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0cf22261-27b0-4ee1-9544-7c52613c1530 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.661783] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 987.661783] env[62521]: value = "task-1318928" [ 987.661783] env[62521]: _type = "Task" [ 987.661783] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.669166] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318928, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.941252] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "refresh_cache-a0646a00-1f1c-4892-bb32-6212c90e9e95" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.941987] env[62521]: DEBUG nova.compute.manager [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Instance network_info: |[{"id": "ff67aa5f-1a3c-478f-bf15-cf0577756a05", "address": "fa:16:3e:65:50:04", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff67aa5f-1a", "ovs_interfaceid": "ff67aa5f-1a3c-478f-bf15-cf0577756a05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.942481] env[62521]: DEBUG oslo_concurrency.lockutils [req-1dab3caa-a5ce-42cc-894b-f06cabe77cdb req-79a9235e-e36f-446e-b9ad-f168ccfcdd8b service nova] Acquired lock "refresh_cache-a0646a00-1f1c-4892-bb32-6212c90e9e95" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.942725] env[62521]: DEBUG nova.network.neutron [req-1dab3caa-a5ce-42cc-894b-f06cabe77cdb req-79a9235e-e36f-446e-b9ad-f168ccfcdd8b service nova] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Refreshing network info cache for port ff67aa5f-1a3c-478f-bf15-cf0577756a05 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 987.944212] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:50:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff67aa5f-1a3c-478f-bf15-cf0577756a05', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.953534] env[62521]: DEBUG oslo.service.loopingcall [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.958293] env[62521]: DEBUG nova.compute.utils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.962047] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.963017] env[62521]: DEBUG nova.compute.manager [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.963227] env[62521]: DEBUG nova.network.neutron [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 987.965568] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d9814d9-c8bb-40b0-9291-dcee17e9f04c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.989934] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.989934] env[62521]: value = "task-1318929" [ 987.989934] env[62521]: _type = "Task" [ 987.989934] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.999845] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.000114] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318929, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.000358] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.035386] env[62521]: DEBUG nova.policy [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3566601952144168b5d400c41026c81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0dfb5f5bd4646ec95b7c6a2d0434175', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 988.091127] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318927, 'name': Rename_Task, 'duration_secs': 0.138232} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.094100] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 988.094528] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6d05509-9f3f-453a-8412-dbdae6f7b331 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.102067] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 988.102067] env[62521]: value = "task-1318930" [ 988.102067] env[62521]: _type = "Task" [ 988.102067] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.110216] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318930, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.138660] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318926, 'name': CloneVM_Task} progress is 95%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.180908] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318928, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.252432] env[62521]: DEBUG nova.network.neutron [req-1dab3caa-a5ce-42cc-894b-f06cabe77cdb req-79a9235e-e36f-446e-b9ad-f168ccfcdd8b service nova] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Updated VIF entry in instance network info cache for port ff67aa5f-1a3c-478f-bf15-cf0577756a05. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 988.252860] env[62521]: DEBUG nova.network.neutron [req-1dab3caa-a5ce-42cc-894b-f06cabe77cdb req-79a9235e-e36f-446e-b9ad-f168ccfcdd8b service nova] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Updating instance_info_cache with network_info: [{"id": "ff67aa5f-1a3c-478f-bf15-cf0577756a05", "address": "fa:16:3e:65:50:04", "network": {"id": "17f39e70-2318-4eec-ba54-622fe1c2232d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1602646973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e6baab1556e4b00b07ab554664b4445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff67aa5f-1a", "ovs_interfaceid": "ff67aa5f-1a3c-478f-bf15-cf0577756a05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.324623] env[62521]: DEBUG nova.network.neutron [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Successfully created port: 56ec9ce6-ead9-4a28-b282-1bfeba30bdf6 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 988.467357] env[62521]: DEBUG nova.compute.manager [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 988.508024] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318929, 'name': CreateVM_Task, 'duration_secs': 0.442326} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.513517] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.513892] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.513991] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 988.518484] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.518661] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.519025] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.519778] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e749c780-00f3-4da5-8260-6ed2906fd64e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.527743] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 988.527743] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5266c5d3-9629-4ed9-1014-c67b314852a5" [ 988.527743] env[62521]: _type = "Task" [ 988.527743] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.542658] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5266c5d3-9629-4ed9-1014-c67b314852a5, 'name': SearchDatastore_Task, 'duration_secs': 0.011373} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.543415] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.543415] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.543558] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.543932] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.543932] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.544172] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5030579e-e37a-43ab-a8b1-ed17da0cda47 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.555510] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.555707] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 988.556492] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3216fd4d-0f18-4e67-bf3b-7424a199d204 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.567020] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 988.567020] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a36c05-1488-7fe4-7c70-1bf2fb9d5254" [ 988.567020] env[62521]: _type = "Task" [ 988.567020] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.573096] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a36c05-1488-7fe4-7c70-1bf2fb9d5254, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.611258] env[62521]: DEBUG oslo_vmware.api [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1318930, 'name': PowerOnVM_Task, 'duration_secs': 0.505859} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.611551] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 988.611763] env[62521]: INFO nova.compute.manager [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Took 7.89 seconds to spawn the instance on the hypervisor. [ 988.611947] env[62521]: DEBUG nova.compute.manager [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.612746] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e66dd63-a156-427f-9807-53a8dd1ba6c6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.638315] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318926, 'name': CloneVM_Task, 'duration_secs': 1.336923} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.638582] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Created linked-clone VM from snapshot [ 988.639313] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5011f487-ebe7-4c97-bd2c-990a5d388011 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.646147] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Uploading image 2d8e21ca-6710-453b-83a9-3f2485192c5d {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 988.673319] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318928, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56134} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.677089] env[62521]: DEBUG oslo_vmware.rw_handles [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 988.677089] env[62521]: value = "vm-282159" [ 988.677089] env[62521]: _type = "VirtualMachine" [ 988.677089] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 988.677381] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] a68d4cb7-bfad-48f6-8aeb-bca04248a0d7/a68d4cb7-bfad-48f6-8aeb-bca04248a0d7.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 988.677602] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.677839] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-842e6e91-5e42-4682-8c36-caa78996d34c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.679972] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ecc96df-6d7a-4321-905c-e6c617ece506 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.685240] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 988.685240] env[62521]: value = "task-1318931" [ 988.685240] env[62521]: _type = "Task" [ 988.685240] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.686418] env[62521]: DEBUG oslo_vmware.rw_handles [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lease: (returnval){ [ 988.686418] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522aa3fb-114e-3e43-e4a3-edca1d64a525" [ 988.686418] env[62521]: _type = "HttpNfcLease" [ 988.686418] env[62521]: } obtained for exporting VM: (result){ [ 988.686418] env[62521]: value = "vm-282159" [ 988.686418] env[62521]: _type = "VirtualMachine" [ 988.686418] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 988.686697] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the lease: (returnval){ [ 988.686697] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522aa3fb-114e-3e43-e4a3-edca1d64a525" [ 988.686697] env[62521]: _type = "HttpNfcLease" [ 988.686697] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 988.702570] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318931, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.704137] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 988.704137] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522aa3fb-114e-3e43-e4a3-edca1d64a525" [ 988.704137] env[62521]: _type = "HttpNfcLease" [ 988.704137] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 988.704422] env[62521]: DEBUG oslo_vmware.rw_handles [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 988.704422] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522aa3fb-114e-3e43-e4a3-edca1d64a525" [ 988.704422] env[62521]: _type = "HttpNfcLease" [ 988.704422] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 988.705203] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db84270-504a-4aea-81b5-3d161701ebe5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.714709] env[62521]: DEBUG oslo_vmware.rw_handles [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52725636-997e-846b-20ed-1c47dd190b15/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 988.714859] env[62521]: DEBUG oslo_vmware.rw_handles [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52725636-997e-846b-20ed-1c47dd190b15/disk-0.vmdk for reading. {{(pid=62521) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 988.791607] env[62521]: DEBUG oslo_concurrency.lockutils [req-1dab3caa-a5ce-42cc-894b-f06cabe77cdb req-79a9235e-e36f-446e-b9ad-f168ccfcdd8b service nova] Releasing lock "refresh_cache-a0646a00-1f1c-4892-bb32-6212c90e9e95" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.794326] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e99cb4-851b-4938-b056-0f4def33f56a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.805942] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8009c7-2b3a-46a1-b92f-a603fbe42bfa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.840497] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd711fa-f7f3-45fe-a7ac-35cc5fe585f2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.842707] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-07b3ad36-c0b6-46a1-8856-656e58da6073 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.850150] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f98b3e-36a5-4d66-83ef-28fc415aa79d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.864772] env[62521]: DEBUG nova.compute.provider_tree [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.076991] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a36c05-1488-7fe4-7c70-1bf2fb9d5254, 'name': SearchDatastore_Task, 'duration_secs': 0.009495} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.077916] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78eed752-08fa-4e58-97b3-2399bead9029 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.083997] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 989.083997] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52dce4a4-2069-9347-c590-b15cf94dd28a" [ 989.083997] env[62521]: _type = "Task" [ 989.083997] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.091685] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52dce4a4-2069-9347-c590-b15cf94dd28a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.131327] env[62521]: INFO nova.compute.manager [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Took 18.21 seconds to build instance. [ 989.196380] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318931, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127026} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.196667] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.197602] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16cdb1a-6150-4992-8883-4e1e8379a5ab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.220363] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] a68d4cb7-bfad-48f6-8aeb-bca04248a0d7/a68d4cb7-bfad-48f6-8aeb-bca04248a0d7.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.221082] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-250a1e06-5962-419f-9a4c-438ec59d073d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.240761] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 989.240761] env[62521]: value = "task-1318933" [ 989.240761] env[62521]: _type = "Task" [ 989.240761] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.249817] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318933, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.368551] env[62521]: DEBUG nova.scheduler.client.report [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.477662] env[62521]: DEBUG nova.compute.manager [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 989.501412] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 989.501784] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 989.502020] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 989.502327] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 989.502618] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 989.502807] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 989.503115] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 989.503447] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 989.503652] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 989.503834] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 989.504084] env[62521]: DEBUG nova.virt.hardware [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.505034] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f2fc05-3dd4-414e-86a4-99ab55eb00d0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.513144] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb8fc4d-55f4-4a0a-b767-fa7b09a59d34 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.596138] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52dce4a4-2069-9347-c590-b15cf94dd28a, 'name': SearchDatastore_Task, 'duration_secs': 0.01188} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.596519] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.597212] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] a0646a00-1f1c-4892-bb32-6212c90e9e95/a0646a00-1f1c-4892-bb32-6212c90e9e95.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 989.597468] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9121d1a-3a35-4199-9480-26def9e814ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.606373] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 989.606373] env[62521]: value = "task-1318934" [ 989.606373] env[62521]: _type = "Task" [ 989.606373] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.614931] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318934, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.633699] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b8a31755-d3f8-4693-ab93-6fa9a352b8be tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.719s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.750781] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318933, 'name': ReconfigVM_Task, 'duration_secs': 0.363195} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.751958] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Reconfigured VM instance instance-00000058 to attach disk [datastore1] a68d4cb7-bfad-48f6-8aeb-bca04248a0d7/a68d4cb7-bfad-48f6-8aeb-bca04248a0d7.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.751958] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c8b115c-1d86-4c50-b709-1deedef3f90a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.757376] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 989.757376] env[62521]: value = "task-1318935" [ 989.757376] env[62521]: _type = "Task" [ 989.757376] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.765422] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318935, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.874196] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.874982] env[62521]: DEBUG nova.compute.manager [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 989.877930] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.441s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.879435] env[62521]: INFO nova.compute.claims [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.122018] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318934, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.267281] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318935, 'name': Rename_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.319512] env[62521]: DEBUG nova.compute.manager [req-ddb79097-84af-42f7-9ba0-6a136b0f7b9a req-3366238e-bac4-4a67-a9c7-775d76694328 service nova] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Received event network-vif-plugged-56ec9ce6-ead9-4a28-b282-1bfeba30bdf6 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.321535] env[62521]: DEBUG oslo_concurrency.lockutils [req-ddb79097-84af-42f7-9ba0-6a136b0f7b9a req-3366238e-bac4-4a67-a9c7-775d76694328 service nova] Acquiring lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.321535] env[62521]: DEBUG oslo_concurrency.lockutils [req-ddb79097-84af-42f7-9ba0-6a136b0f7b9a req-3366238e-bac4-4a67-a9c7-775d76694328 service nova] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.322095] env[62521]: DEBUG oslo_concurrency.lockutils [req-ddb79097-84af-42f7-9ba0-6a136b0f7b9a req-3366238e-bac4-4a67-a9c7-775d76694328 service nova] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.322778] env[62521]: DEBUG nova.compute.manager [req-ddb79097-84af-42f7-9ba0-6a136b0f7b9a req-3366238e-bac4-4a67-a9c7-775d76694328 service nova] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] No waiting events found dispatching network-vif-plugged-56ec9ce6-ead9-4a28-b282-1bfeba30bdf6 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 990.323733] env[62521]: WARNING nova.compute.manager [req-ddb79097-84af-42f7-9ba0-6a136b0f7b9a req-3366238e-bac4-4a67-a9c7-775d76694328 service nova] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Received unexpected event network-vif-plugged-56ec9ce6-ead9-4a28-b282-1bfeba30bdf6 for instance with vm_state building and task_state spawning. [ 990.384476] env[62521]: DEBUG nova.compute.utils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 990.388898] env[62521]: DEBUG nova.compute.manager [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 990.389499] env[62521]: DEBUG nova.network.neutron [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 990.433559] env[62521]: DEBUG nova.policy [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76d85009c2c1416b9ea049003e6d96a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c533bd4917b4466e9c5c76a9caf8e807', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 990.554738] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.554738] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquired lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.554738] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Forcefully refreshing network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 990.602321] env[62521]: DEBUG nova.compute.manager [req-c12ec6d4-97e5-4d0e-b2d2-1a16ad337325 req-2c6aa01e-728b-47a9-9d5e-bcb8c18ecf59 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Received event network-changed-dee03f5c-2442-438b-8786-6f3e0c35f8ab {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.602321] env[62521]: DEBUG nova.compute.manager [req-c12ec6d4-97e5-4d0e-b2d2-1a16ad337325 req-2c6aa01e-728b-47a9-9d5e-bcb8c18ecf59 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Refreshing instance network info cache due to event network-changed-dee03f5c-2442-438b-8786-6f3e0c35f8ab. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 990.602321] env[62521]: DEBUG oslo_concurrency.lockutils [req-c12ec6d4-97e5-4d0e-b2d2-1a16ad337325 req-2c6aa01e-728b-47a9-9d5e-bcb8c18ecf59 service nova] Acquiring lock "refresh_cache-7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.602321] env[62521]: DEBUG oslo_concurrency.lockutils [req-c12ec6d4-97e5-4d0e-b2d2-1a16ad337325 req-2c6aa01e-728b-47a9-9d5e-bcb8c18ecf59 service nova] Acquired lock "refresh_cache-7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.602321] env[62521]: DEBUG nova.network.neutron [req-c12ec6d4-97e5-4d0e-b2d2-1a16ad337325 req-2c6aa01e-728b-47a9-9d5e-bcb8c18ecf59 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Refreshing network info cache for port dee03f5c-2442-438b-8786-6f3e0c35f8ab {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.621961] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318934, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602125} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.622662] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] a0646a00-1f1c-4892-bb32-6212c90e9e95/a0646a00-1f1c-4892-bb32-6212c90e9e95.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 990.623169] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 990.623802] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90b5278b-a04a-4ec8-9453-c4c2f07b0846 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.631217] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 990.631217] env[62521]: value = "task-1318936" [ 990.631217] env[62521]: _type = "Task" [ 990.631217] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.640967] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318936, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.694978] env[62521]: DEBUG nova.network.neutron [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Successfully created port: b4b87574-5fdf-40a2-b2ac-a97129208f8a {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 990.768345] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318935, 'name': Rename_Task, 'duration_secs': 0.510922} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.768887] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 990.769216] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d717a89d-c858-438d-80b0-e576fd24f1af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.775579] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 990.775579] env[62521]: value = "task-1318937" [ 990.775579] env[62521]: _type = "Task" [ 990.775579] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.783206] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318937, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.890283] env[62521]: DEBUG nova.compute.manager [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 990.914160] env[62521]: DEBUG nova.network.neutron [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Successfully updated port: 56ec9ce6-ead9-4a28-b282-1bfeba30bdf6 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.080249] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 991.111704] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9613e4c6-54af-47fa-b844-a63c1a705879 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.124206] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4dc06c-34d8-4739-a7b3-dfced4327cae {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.169588] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba02c989-98d6-4d02-ae17-19c4a9f487a5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.170309] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318936, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088078} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.174015] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.175577] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e18d08d-0d9a-4379-972e-6e627c5fccf6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.179799] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a697bef2-416a-4775-af38-357f3c08aee8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.212790] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] a0646a00-1f1c-4892-bb32-6212c90e9e95/a0646a00-1f1c-4892-bb32-6212c90e9e95.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.221227] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f47b0502-f665-474b-a215-97fa51add5f7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.236676] env[62521]: DEBUG nova.compute.provider_tree [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.244504] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 991.244504] env[62521]: value = "task-1318938" [ 991.244504] env[62521]: _type = "Task" [ 991.244504] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.257122] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318938, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.291594] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318937, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.416991] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "refresh_cache-7cdeed76-e5b3-4959-b6a6-610d23233e1b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.417207] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "refresh_cache-7cdeed76-e5b3-4959-b6a6-610d23233e1b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.417388] env[62521]: DEBUG nova.network.neutron [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 991.637536] env[62521]: DEBUG nova.network.neutron [req-c12ec6d4-97e5-4d0e-b2d2-1a16ad337325 req-2c6aa01e-728b-47a9-9d5e-bcb8c18ecf59 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Updated VIF entry in instance network info cache for port dee03f5c-2442-438b-8786-6f3e0c35f8ab. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.637921] env[62521]: DEBUG nova.network.neutron [req-c12ec6d4-97e5-4d0e-b2d2-1a16ad337325 req-2c6aa01e-728b-47a9-9d5e-bcb8c18ecf59 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Updating instance_info_cache with network_info: [{"id": "dee03f5c-2442-438b-8786-6f3e0c35f8ab", "address": "fa:16:3e:64:68:af", "network": {"id": "a088cee4-0265-4dd0-a57a-bc9e09a1da5b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1586004199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f029ddf5cd8448caad5ec23775b6503", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdee03f5c-24", "ovs_interfaceid": "dee03f5c-2442-438b-8786-6f3e0c35f8ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.689235] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.742066] env[62521]: DEBUG nova.scheduler.client.report [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.755453] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318938, 'name': ReconfigVM_Task, 'duration_secs': 0.394387} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.755703] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Reconfigured VM instance instance-00000059 to attach disk [datastore1] a0646a00-1f1c-4892-bb32-6212c90e9e95/a0646a00-1f1c-4892-bb32-6212c90e9e95.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.756949] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e63ee82-7d99-4a81-bd52-8037308102e4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.764230] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 991.764230] env[62521]: value = "task-1318939" [ 991.764230] env[62521]: _type = "Task" [ 991.764230] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.772663] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318939, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.788295] env[62521]: DEBUG oslo_vmware.api [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1318937, 'name': PowerOnVM_Task, 'duration_secs': 0.538546} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.788464] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 991.788603] env[62521]: INFO nova.compute.manager [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Took 8.75 seconds to spawn the instance on the hypervisor. [ 991.788789] env[62521]: DEBUG nova.compute.manager [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.789605] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a5b004-9c43-4aff-9b3d-53d38621bce1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.901440] env[62521]: DEBUG nova.compute.manager [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 991.931581] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.931917] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.932206] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.932433] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.932636] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.932836] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.933114] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.933313] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.933545] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.933754] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.933966] env[62521]: DEBUG nova.virt.hardware [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.934982] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e155d7e-2e03-4b4f-bac6-ce8f61289bdc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.943619] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ae7a83-8c31-48d4-b90c-1f5070d5b6e6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.959295] env[62521]: DEBUG nova.network.neutron [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 992.096564] env[62521]: DEBUG nova.network.neutron [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Updating instance_info_cache with network_info: [{"id": "56ec9ce6-ead9-4a28-b282-1bfeba30bdf6", "address": "fa:16:3e:da:38:2a", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56ec9ce6-ea", "ovs_interfaceid": "56ec9ce6-ead9-4a28-b282-1bfeba30bdf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.112064] env[62521]: DEBUG nova.compute.manager [req-86685283-ba4d-4cf6-b6d7-a7444f09ee68 req-f307873f-af33-40bc-bdeb-7a0875e0542b service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Received event network-vif-plugged-b4b87574-5fdf-40a2-b2ac-a97129208f8a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.112397] env[62521]: DEBUG oslo_concurrency.lockutils [req-86685283-ba4d-4cf6-b6d7-a7444f09ee68 req-f307873f-af33-40bc-bdeb-7a0875e0542b service nova] Acquiring lock "69c5ff30-259d-4067-b5e2-ab727009d8ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.112582] env[62521]: DEBUG oslo_concurrency.lockutils [req-86685283-ba4d-4cf6-b6d7-a7444f09ee68 req-f307873f-af33-40bc-bdeb-7a0875e0542b service nova] Lock "69c5ff30-259d-4067-b5e2-ab727009d8ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.112684] env[62521]: DEBUG oslo_concurrency.lockutils [req-86685283-ba4d-4cf6-b6d7-a7444f09ee68 req-f307873f-af33-40bc-bdeb-7a0875e0542b service nova] Lock "69c5ff30-259d-4067-b5e2-ab727009d8ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.112865] env[62521]: DEBUG nova.compute.manager [req-86685283-ba4d-4cf6-b6d7-a7444f09ee68 req-f307873f-af33-40bc-bdeb-7a0875e0542b service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] No waiting events found dispatching network-vif-plugged-b4b87574-5fdf-40a2-b2ac-a97129208f8a {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.113030] env[62521]: WARNING nova.compute.manager [req-86685283-ba4d-4cf6-b6d7-a7444f09ee68 req-f307873f-af33-40bc-bdeb-7a0875e0542b service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Received unexpected event network-vif-plugged-b4b87574-5fdf-40a2-b2ac-a97129208f8a for instance with vm_state building and task_state spawning. [ 992.143088] env[62521]: DEBUG oslo_concurrency.lockutils [req-c12ec6d4-97e5-4d0e-b2d2-1a16ad337325 req-2c6aa01e-728b-47a9-9d5e-bcb8c18ecf59 service nova] Releasing lock "refresh_cache-7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.191755] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Releasing lock "refresh_cache-bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.192193] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Updated the network info_cache for instance {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 992.192440] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 992.192608] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 992.192774] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 992.192902] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 992.193068] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 992.193204] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 992.193364] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 992.193476] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 992.195709] env[62521]: DEBUG nova.network.neutron [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Successfully updated port: b4b87574-5fdf-40a2-b2ac-a97129208f8a {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 992.251543] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.251901] env[62521]: DEBUG nova.compute.manager [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 992.254648] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.330s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.254995] env[62521]: DEBUG nova.objects.instance [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lazy-loading 'resources' on Instance uuid bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.275188] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318939, 'name': Rename_Task, 'duration_secs': 0.151743} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.276028] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.276028] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e1a798c1-4d93-4adc-9b9b-aa10943ea0ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.286477] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 992.286477] env[62521]: value = "task-1318940" [ 992.286477] env[62521]: _type = "Task" [ 992.286477] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.292785] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318940, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.307034] env[62521]: INFO nova.compute.manager [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Took 18.83 seconds to build instance. [ 992.602017] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "refresh_cache-7cdeed76-e5b3-4959-b6a6-610d23233e1b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.602017] env[62521]: DEBUG nova.compute.manager [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Instance network_info: |[{"id": "56ec9ce6-ead9-4a28-b282-1bfeba30bdf6", "address": "fa:16:3e:da:38:2a", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56ec9ce6-ea", "ovs_interfaceid": "56ec9ce6-ead9-4a28-b282-1bfeba30bdf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 992.602017] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:38:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '304be4f7-4e36-4468-9ef4-e457341cef18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '56ec9ce6-ead9-4a28-b282-1bfeba30bdf6', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.609563] env[62521]: DEBUG oslo.service.loopingcall [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.609946] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 992.610330] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c1c30fc-53cc-4472-a231-367877f871a4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.632802] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.632802] env[62521]: value = "task-1318941" [ 992.632802] env[62521]: _type = "Task" [ 992.632802] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.641800] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318941, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.696747] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.698817] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "refresh_cache-69c5ff30-259d-4067-b5e2-ab727009d8ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.699202] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "refresh_cache-69c5ff30-259d-4067-b5e2-ab727009d8ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.700583] env[62521]: DEBUG nova.network.neutron [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 992.760023] env[62521]: DEBUG nova.compute.utils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 992.762574] env[62521]: DEBUG nova.compute.manager [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 992.762949] env[62521]: DEBUG nova.network.neutron [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 992.801505] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318940, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.811368] env[62521]: DEBUG oslo_concurrency.lockutils [None req-fd05f5b2-c57d-4db4-8a40-81f8b448c2c6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.344s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.815810] env[62521]: DEBUG nova.policy [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e7c7fb132924e8d8ae25f4cdda9f63c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f513dc5025974607a90e67621154bccd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 992.990025] env[62521]: DEBUG nova.compute.manager [req-8df0519d-c543-4ad5-8bc3-de3406ccffcc req-4f6dd8cf-8a0e-470d-b84f-639e7995fbf9 service nova] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Received event network-changed-56ec9ce6-ead9-4a28-b282-1bfeba30bdf6 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.990190] env[62521]: DEBUG nova.compute.manager [req-8df0519d-c543-4ad5-8bc3-de3406ccffcc req-4f6dd8cf-8a0e-470d-b84f-639e7995fbf9 service nova] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Refreshing instance network info cache due to event network-changed-56ec9ce6-ead9-4a28-b282-1bfeba30bdf6. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 992.990336] env[62521]: DEBUG oslo_concurrency.lockutils [req-8df0519d-c543-4ad5-8bc3-de3406ccffcc req-4f6dd8cf-8a0e-470d-b84f-639e7995fbf9 service nova] Acquiring lock "refresh_cache-7cdeed76-e5b3-4959-b6a6-610d23233e1b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.990483] env[62521]: DEBUG oslo_concurrency.lockutils [req-8df0519d-c543-4ad5-8bc3-de3406ccffcc req-4f6dd8cf-8a0e-470d-b84f-639e7995fbf9 service nova] Acquired lock "refresh_cache-7cdeed76-e5b3-4959-b6a6-610d23233e1b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.990652] env[62521]: DEBUG nova.network.neutron [req-8df0519d-c543-4ad5-8bc3-de3406ccffcc req-4f6dd8cf-8a0e-470d-b84f-639e7995fbf9 service nova] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Refreshing network info cache for port 56ec9ce6-ead9-4a28-b282-1bfeba30bdf6 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.065160] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156917a9-8d35-4e5c-aff8-5ae6a7d05b8a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.073427] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f42859-ac57-4303-b6e6-0e83b03ecf4a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.115120] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf47329-9520-4b2f-a75e-664982a13095 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.118569] env[62521]: DEBUG nova.network.neutron [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Successfully created port: 75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 993.126937] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a3aa7c-eca2-4ef6-a5a8-20261d883ddd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.151167] env[62521]: DEBUG nova.compute.provider_tree [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.158464] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318941, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.253849] env[62521]: DEBUG nova.network.neutron [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 993.267810] env[62521]: DEBUG nova.compute.manager [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 993.295236] env[62521]: DEBUG oslo_vmware.api [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318940, 'name': PowerOnVM_Task, 'duration_secs': 0.51709} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.295539] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.295812] env[62521]: INFO nova.compute.manager [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Took 7.94 seconds to spawn the instance on the hypervisor. [ 993.295913] env[62521]: DEBUG nova.compute.manager [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.296739] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbd95bf-983d-41f0-9823-1673067a16ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.495771] env[62521]: DEBUG nova.network.neutron [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Updating instance_info_cache with network_info: [{"id": "b4b87574-5fdf-40a2-b2ac-a97129208f8a", "address": "fa:16:3e:0a:a7:df", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4b87574-5f", "ovs_interfaceid": "b4b87574-5fdf-40a2-b2ac-a97129208f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.644709] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318941, 'name': CreateVM_Task, 'duration_secs': 0.608427} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.645036] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 993.645563] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.645737] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.646070] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.648704] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5afc6c89-d03e-497d-821d-c70bd5cf2ce6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.653330] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 993.653330] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5248a7e0-9582-bfde-8540-bcc4b1c5571e" [ 993.653330] env[62521]: _type = "Task" [ 993.653330] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.661585] env[62521]: DEBUG nova.scheduler.client.report [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.665032] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5248a7e0-9582-bfde-8540-bcc4b1c5571e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.750962] env[62521]: DEBUG nova.network.neutron [req-8df0519d-c543-4ad5-8bc3-de3406ccffcc req-4f6dd8cf-8a0e-470d-b84f-639e7995fbf9 service nova] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Updated VIF entry in instance network info cache for port 56ec9ce6-ead9-4a28-b282-1bfeba30bdf6. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 993.751361] env[62521]: DEBUG nova.network.neutron [req-8df0519d-c543-4ad5-8bc3-de3406ccffcc req-4f6dd8cf-8a0e-470d-b84f-639e7995fbf9 service nova] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Updating instance_info_cache with network_info: [{"id": "56ec9ce6-ead9-4a28-b282-1bfeba30bdf6", "address": "fa:16:3e:da:38:2a", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56ec9ce6-ea", "ovs_interfaceid": "56ec9ce6-ead9-4a28-b282-1bfeba30bdf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.817362] env[62521]: INFO nova.compute.manager [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Took 18.42 seconds to build instance. [ 994.000090] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "refresh_cache-69c5ff30-259d-4067-b5e2-ab727009d8ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.000298] env[62521]: DEBUG nova.compute.manager [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Instance network_info: |[{"id": "b4b87574-5fdf-40a2-b2ac-a97129208f8a", "address": "fa:16:3e:0a:a7:df", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4b87574-5f", "ovs_interfaceid": "b4b87574-5fdf-40a2-b2ac-a97129208f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 994.000839] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:a7:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '84aee122-f630-43c5-9cc1-3a38d3819c82', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4b87574-5fdf-40a2-b2ac-a97129208f8a', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.010367] env[62521]: DEBUG oslo.service.loopingcall [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.010697] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.011041] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-037d755a-ff29-4257-9190-0b7758f1ebc8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.036965] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.036965] env[62521]: value = "task-1318942" [ 994.036965] env[62521]: _type = "Task" [ 994.036965] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.046439] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318942, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.142776] env[62521]: DEBUG nova.compute.manager [req-6f70bfb3-1a2b-465d-9721-09de34d0502b req-97bd187e-2bfd-4554-b9fa-115a8ea773a6 service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Received event network-changed-b4b87574-5fdf-40a2-b2ac-a97129208f8a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.142990] env[62521]: DEBUG nova.compute.manager [req-6f70bfb3-1a2b-465d-9721-09de34d0502b req-97bd187e-2bfd-4554-b9fa-115a8ea773a6 service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Refreshing instance network info cache due to event network-changed-b4b87574-5fdf-40a2-b2ac-a97129208f8a. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 994.143225] env[62521]: DEBUG oslo_concurrency.lockutils [req-6f70bfb3-1a2b-465d-9721-09de34d0502b req-97bd187e-2bfd-4554-b9fa-115a8ea773a6 service nova] Acquiring lock "refresh_cache-69c5ff30-259d-4067-b5e2-ab727009d8ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.143376] env[62521]: DEBUG oslo_concurrency.lockutils [req-6f70bfb3-1a2b-465d-9721-09de34d0502b req-97bd187e-2bfd-4554-b9fa-115a8ea773a6 service nova] Acquired lock "refresh_cache-69c5ff30-259d-4067-b5e2-ab727009d8ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.143540] env[62521]: DEBUG nova.network.neutron [req-6f70bfb3-1a2b-465d-9721-09de34d0502b req-97bd187e-2bfd-4554-b9fa-115a8ea773a6 service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Refreshing network info cache for port b4b87574-5fdf-40a2-b2ac-a97129208f8a {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.166190] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5248a7e0-9582-bfde-8540-bcc4b1c5571e, 'name': SearchDatastore_Task, 'duration_secs': 0.025504} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.167491] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.913s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.169722] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.170302] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.170686] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.170954] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.171321] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.172038] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.640s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.175259] env[62521]: INFO nova.compute.claims [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.176443] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a6e860f-ef02-4404-b576-f5b4558563d0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.193797] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.194283] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 994.195235] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55a266cc-a1aa-463d-ade4-84dbad17d495 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.202677] env[62521]: INFO nova.scheduler.client.report [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Deleted allocations for instance bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0 [ 994.203236] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 994.203236] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523e01f7-ae96-efda-9de5-37edf2092f55" [ 994.203236] env[62521]: _type = "Task" [ 994.203236] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.218800] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523e01f7-ae96-efda-9de5-37edf2092f55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.253832] env[62521]: DEBUG oslo_concurrency.lockutils [req-8df0519d-c543-4ad5-8bc3-de3406ccffcc req-4f6dd8cf-8a0e-470d-b84f-639e7995fbf9 service nova] Releasing lock "refresh_cache-7cdeed76-e5b3-4959-b6a6-610d23233e1b" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.275840] env[62521]: DEBUG nova.compute.manager [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 994.307180] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 994.307457] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 994.307624] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 994.307822] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 994.308029] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 994.308202] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 994.308441] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 994.309033] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 994.309033] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 994.309033] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 994.309268] env[62521]: DEBUG nova.virt.hardware [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 994.310018] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be2f500-f62c-4bd8-abf2-27d01c1e42b1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.317927] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e2f25d-fde8-446c-8daf-f02c7bb31305 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.322182] env[62521]: DEBUG oslo_concurrency.lockutils [None req-05942532-e5b6-4df3-9731-73fecaa24b38 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "a0646a00-1f1c-4892-bb32-6212c90e9e95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.936s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.545060] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318942, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.655232] env[62521]: DEBUG nova.network.neutron [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Successfully updated port: 75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 994.713834] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9846b128-b28b-436f-8f8a-d0b7824b96a0 tempest-ServersNegativeTestJSON-997986063 tempest-ServersNegativeTestJSON-997986063-project-member] Lock "bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.647s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.719795] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523e01f7-ae96-efda-9de5-37edf2092f55, 'name': SearchDatastore_Task, 'duration_secs': 0.013866} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.720273] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b06dfc0b-734f-4d0c-8b91-af658c77a151 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.725599] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 994.725599] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52873fb3-021a-5fe1-431e-1cda754c8f4f" [ 994.725599] env[62521]: _type = "Task" [ 994.725599] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.734182] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52873fb3-021a-5fe1-431e-1cda754c8f4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.947237] env[62521]: DEBUG nova.network.neutron [req-6f70bfb3-1a2b-465d-9721-09de34d0502b req-97bd187e-2bfd-4554-b9fa-115a8ea773a6 service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Updated VIF entry in instance network info cache for port b4b87574-5fdf-40a2-b2ac-a97129208f8a. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 994.947672] env[62521]: DEBUG nova.network.neutron [req-6f70bfb3-1a2b-465d-9721-09de34d0502b req-97bd187e-2bfd-4554-b9fa-115a8ea773a6 service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Updating instance_info_cache with network_info: [{"id": "b4b87574-5fdf-40a2-b2ac-a97129208f8a", "address": "fa:16:3e:0a:a7:df", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4b87574-5f", "ovs_interfaceid": "b4b87574-5fdf-40a2-b2ac-a97129208f8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.018343] env[62521]: DEBUG nova.compute.manager [req-1d0a73e6-13ed-4bb3-bee4-16f3f0742151 req-7455f227-91be-471c-a7bc-22dd395cee80 service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Received event network-changed-add56f48-9368-4591-b8b6-cc7be1121d28 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.018658] env[62521]: DEBUG nova.compute.manager [req-1d0a73e6-13ed-4bb3-bee4-16f3f0742151 req-7455f227-91be-471c-a7bc-22dd395cee80 service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Refreshing instance network info cache due to event network-changed-add56f48-9368-4591-b8b6-cc7be1121d28. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.018972] env[62521]: DEBUG oslo_concurrency.lockutils [req-1d0a73e6-13ed-4bb3-bee4-16f3f0742151 req-7455f227-91be-471c-a7bc-22dd395cee80 service nova] Acquiring lock "refresh_cache-a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.019206] env[62521]: DEBUG oslo_concurrency.lockutils [req-1d0a73e6-13ed-4bb3-bee4-16f3f0742151 req-7455f227-91be-471c-a7bc-22dd395cee80 service nova] Acquired lock "refresh_cache-a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.019444] env[62521]: DEBUG nova.network.neutron [req-1d0a73e6-13ed-4bb3-bee4-16f3f0742151 req-7455f227-91be-471c-a7bc-22dd395cee80 service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Refreshing network info cache for port add56f48-9368-4591-b8b6-cc7be1121d28 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 995.045643] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318942, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.158190] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquiring lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.158527] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquired lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.158576] env[62521]: DEBUG nova.network.neutron [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 995.236667] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52873fb3-021a-5fe1-431e-1cda754c8f4f, 'name': SearchDatastore_Task, 'duration_secs': 0.025347} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.239433] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.239729] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 7cdeed76-e5b3-4959-b6a6-610d23233e1b/7cdeed76-e5b3-4959-b6a6-610d23233e1b.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 995.241190] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49474f31-0daf-4064-9188-c6a163c402d2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.247746] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 995.247746] env[62521]: value = "task-1318943" [ 995.247746] env[62521]: _type = "Task" [ 995.247746] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.259756] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.384144] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c71357-4d4a-4e98-acbc-42183a8b099f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.392193] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bcfcbef-05d3-4dc0-b5af-312b8981e848 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.423774] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871a36e0-52fc-40d8-9209-3552bee9b640 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.432861] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b5da30-d844-4c12-ace8-f4026f62eac8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.451305] env[62521]: DEBUG oslo_concurrency.lockutils [req-6f70bfb3-1a2b-465d-9721-09de34d0502b req-97bd187e-2bfd-4554-b9fa-115a8ea773a6 service nova] Releasing lock "refresh_cache-69c5ff30-259d-4067-b5e2-ab727009d8ad" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.451905] env[62521]: DEBUG nova.compute.provider_tree [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.546541] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318942, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.552751] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "a0646a00-1f1c-4892-bb32-6212c90e9e95" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.552751] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "a0646a00-1f1c-4892-bb32-6212c90e9e95" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.552751] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "a0646a00-1f1c-4892-bb32-6212c90e9e95-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.552751] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "a0646a00-1f1c-4892-bb32-6212c90e9e95-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.552751] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "a0646a00-1f1c-4892-bb32-6212c90e9e95-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.554652] env[62521]: INFO nova.compute.manager [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Terminating instance [ 995.556863] env[62521]: DEBUG nova.compute.manager [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 995.557140] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 995.558027] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e068faf4-57e4-4642-9ea1-b44579f13ef0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.566460] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 995.566617] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ff0e68c-dbf9-4edb-b660-35cb3225ebfa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.572780] env[62521]: DEBUG oslo_vmware.api [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 995.572780] env[62521]: value = "task-1318944" [ 995.572780] env[62521]: _type = "Task" [ 995.572780] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.581607] env[62521]: DEBUG oslo_vmware.api [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318944, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.697492] env[62521]: DEBUG nova.network.neutron [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 995.763020] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318943, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.869951] env[62521]: DEBUG nova.network.neutron [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updating instance_info_cache with network_info: [{"id": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "address": "fa:16:3e:82:b0:68", "network": {"id": "e84e7dd3-8f72-4865-86ba-63a68f5178cf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1633868122-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f513dc5025974607a90e67621154bccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75ffddaa-cb", "ovs_interfaceid": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.956412] env[62521]: DEBUG nova.scheduler.client.report [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.046260] env[62521]: DEBUG nova.network.neutron [req-1d0a73e6-13ed-4bb3-bee4-16f3f0742151 req-7455f227-91be-471c-a7bc-22dd395cee80 service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Updated VIF entry in instance network info cache for port add56f48-9368-4591-b8b6-cc7be1121d28. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 996.046657] env[62521]: DEBUG nova.network.neutron [req-1d0a73e6-13ed-4bb3-bee4-16f3f0742151 req-7455f227-91be-471c-a7bc-22dd395cee80 service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Updating instance_info_cache with network_info: [{"id": "add56f48-9368-4591-b8b6-cc7be1121d28", "address": "fa:16:3e:f2:8e:fd", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadd56f48-93", "ovs_interfaceid": "add56f48-9368-4591-b8b6-cc7be1121d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.051359] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318942, 'name': CreateVM_Task, 'duration_secs': 1.782801} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.051783] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 996.052489] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.052677] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.053187] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.053503] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01e1db5d-da85-4e6c-8dea-64f32f34a749 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.059955] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 996.059955] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524f6dcf-1b1c-68cb-2e91-71ecd88d0457" [ 996.059955] env[62521]: _type = "Task" [ 996.059955] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.073110] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524f6dcf-1b1c-68cb-2e91-71ecd88d0457, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.084893] env[62521]: DEBUG oslo_vmware.api [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318944, 'name': PowerOffVM_Task, 'duration_secs': 0.205321} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.085341] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 996.085987] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 996.085987] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80d20395-2ad9-42a5-8240-c0da27bf2bbc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.153767] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 996.154418] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 996.154418] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleting the datastore file [datastore1] a0646a00-1f1c-4892-bb32-6212c90e9e95 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.154646] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6cf35a39-4b45-4ec6-9d4d-fa2c67ee6a72 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.161511] env[62521]: DEBUG oslo_vmware.api [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for the task: (returnval){ [ 996.161511] env[62521]: value = "task-1318946" [ 996.161511] env[62521]: _type = "Task" [ 996.161511] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.170177] env[62521]: DEBUG oslo_vmware.api [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.179306] env[62521]: DEBUG nova.compute.manager [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Received event network-vif-plugged-75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.179564] env[62521]: DEBUG oslo_concurrency.lockutils [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] Acquiring lock "31f926a8-777d-4882-afd6-1e4dba3f4e11-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.179776] env[62521]: DEBUG oslo_concurrency.lockutils [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] Lock "31f926a8-777d-4882-afd6-1e4dba3f4e11-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.180315] env[62521]: DEBUG oslo_concurrency.lockutils [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] Lock "31f926a8-777d-4882-afd6-1e4dba3f4e11-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.180315] env[62521]: DEBUG nova.compute.manager [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] No waiting events found dispatching network-vif-plugged-75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 996.180450] env[62521]: WARNING nova.compute.manager [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Received unexpected event network-vif-plugged-75ffddaa-cba0-4286-84c8-454e40d5bb4f for instance with vm_state building and task_state spawning. [ 996.180620] env[62521]: DEBUG nova.compute.manager [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Received event network-changed-75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.180844] env[62521]: DEBUG nova.compute.manager [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Refreshing instance network info cache due to event network-changed-75ffddaa-cba0-4286-84c8-454e40d5bb4f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 996.181070] env[62521]: DEBUG oslo_concurrency.lockutils [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] Acquiring lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.263084] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.919749} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.263527] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 7cdeed76-e5b3-4959-b6a6-610d23233e1b/7cdeed76-e5b3-4959-b6a6-610d23233e1b.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 996.263789] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.264483] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de2e4d74-28f6-41de-9d12-5268aaee7a3f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.271764] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 996.271764] env[62521]: value = "task-1318947" [ 996.271764] env[62521]: _type = "Task" [ 996.271764] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.280646] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318947, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.373498] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Releasing lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.373801] env[62521]: DEBUG nova.compute.manager [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Instance network_info: |[{"id": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "address": "fa:16:3e:82:b0:68", "network": {"id": "e84e7dd3-8f72-4865-86ba-63a68f5178cf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1633868122-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f513dc5025974607a90e67621154bccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75ffddaa-cb", "ovs_interfaceid": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 996.374160] env[62521]: DEBUG oslo_concurrency.lockutils [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] Acquired lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.374511] env[62521]: DEBUG nova.network.neutron [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Refreshing network info cache for port 75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 996.376221] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:b0:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ebf80ce9-9885-46ed-ac23-310a98789a95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '75ffddaa-cba0-4286-84c8-454e40d5bb4f', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 996.385991] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Creating folder: Project (f513dc5025974607a90e67621154bccd). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 996.389765] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a3dc3dc-3166-4ce6-b682-9caa5cf9d9fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.403278] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Created folder: Project (f513dc5025974607a90e67621154bccd) in parent group-v282025. [ 996.403700] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Creating folder: Instances. Parent ref: group-v282163. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 996.403997] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2818a0d9-9167-4842-a8d2-39ade07c2b96 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.406884] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.407184] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.418057] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Created folder: Instances in parent group-v282163. [ 996.418057] env[62521]: DEBUG oslo.service.loopingcall [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.418057] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 996.418057] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4af0a873-39de-43e9-b77a-67e7ed618d26 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.442801] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 996.442801] env[62521]: value = "task-1318950" [ 996.442801] env[62521]: _type = "Task" [ 996.442801] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.452274] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318950, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.462805] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.463409] env[62521]: DEBUG nova.compute.manager [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 996.469021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.772s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.469278] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.469389] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 996.470677] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189f13d3-19b2-4969-b68b-68e17122b382 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.479728] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7beb836f-943e-465f-b4f4-f55f476a95f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.498824] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7883b564-f000-400e-a824-7a9b4ba38719 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.506015] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ed962b-a4b2-4aa1-a159-66b8c3d0c35f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.537546] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179784MB free_disk=133GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 996.537871] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.537987] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.553336] env[62521]: DEBUG oslo_concurrency.lockutils [req-1d0a73e6-13ed-4bb3-bee4-16f3f0742151 req-7455f227-91be-471c-a7bc-22dd395cee80 service nova] Releasing lock "refresh_cache-a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.572021] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524f6dcf-1b1c-68cb-2e91-71ecd88d0457, 'name': SearchDatastore_Task, 'duration_secs': 0.062448} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.572021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.572021] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.572021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.572021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.572021] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.574063] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5b55552-9187-4227-ac83-988e4f4dcd15 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.583487] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.587021] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 996.587021] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ad502ac-9337-48d8-96a4-96fdbec9237a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.591150] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 996.591150] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e83fe-3a7f-2c5a-2856-75f5c36883c3" [ 996.591150] env[62521]: _type = "Task" [ 996.591150] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.602485] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e83fe-3a7f-2c5a-2856-75f5c36883c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.671520] env[62521]: DEBUG oslo_vmware.api [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Task: {'id': task-1318946, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.429237} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.674354] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.674354] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 996.674522] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 996.674699] env[62521]: INFO nova.compute.manager [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Took 1.12 seconds to destroy the instance on the hypervisor. [ 996.674984] env[62521]: DEBUG oslo.service.loopingcall [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.675197] env[62521]: DEBUG nova.compute.manager [-] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 996.675291] env[62521]: DEBUG nova.network.neutron [-] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.750530] env[62521]: DEBUG nova.network.neutron [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updated VIF entry in instance network info cache for port 75ffddaa-cba0-4286-84c8-454e40d5bb4f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 996.750905] env[62521]: DEBUG nova.network.neutron [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updating instance_info_cache with network_info: [{"id": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "address": "fa:16:3e:82:b0:68", "network": {"id": "e84e7dd3-8f72-4865-86ba-63a68f5178cf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1633868122-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f513dc5025974607a90e67621154bccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75ffddaa-cb", "ovs_interfaceid": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.781721] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318947, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094331} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.783992] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.784853] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c45a7b-e7ce-4a72-bbeb-5fa31acc5ac0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.808008] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 7cdeed76-e5b3-4959-b6a6-610d23233e1b/7cdeed76-e5b3-4959-b6a6-610d23233e1b.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.808404] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2973d156-eece-4e98-98f0-2144b5a3c988 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.827552] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 996.827552] env[62521]: value = "task-1318951" [ 996.827552] env[62521]: _type = "Task" [ 996.827552] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.837094] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.911030] env[62521]: INFO nova.compute.manager [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Detaching volume a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c [ 996.953994] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318950, 'name': CreateVM_Task, 'duration_secs': 0.407295} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.954187] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 996.954951] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.955159] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.955478] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.959605] env[62521]: INFO nova.virt.block_device [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Attempting to driver detach volume a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c from mountpoint /dev/sdb [ 996.959605] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 996.959605] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282136', 'volume_id': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'name': 'volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '73bc1cf1-7ee1-4a9b-8214-2dc752b700ad', 'attached_at': '', 'detached_at': '', 'volume_id': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'serial': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 996.959605] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfc26538-a32e-461c-bc27-fd690afe0766 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.959892] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c47735-0968-4fb1-921a-e54421c570ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.966186] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 996.966186] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52745248-4811-d339-e6b4-6216e1d29263" [ 996.966186] env[62521]: _type = "Task" [ 996.966186] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.984684] env[62521]: DEBUG nova.compute.utils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.990337] env[62521]: DEBUG nova.compute.manager [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 996.990523] env[62521]: DEBUG nova.network.neutron [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 996.993174] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4747755-a82c-40b2-b391-535b9fb5d4c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.005824] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52745248-4811-d339-e6b4-6216e1d29263, 'name': SearchDatastore_Task, 'duration_secs': 0.017742} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.005824] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.005824] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 997.005824] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.005824] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd374aa-73fd-41e4-a736-05ed65790a0b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.033036] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311f5766-5630-45b2-8661-b2ab031919f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.056215] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] The volume has not been displaced from its original location: [datastore2] volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c/volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 997.061617] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Reconfiguring VM instance instance-00000042 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 997.063299] env[62521]: DEBUG nova.policy [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a0e35b0e4574ed9afc4549d30daa210', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '31c788848b0d47478564e53066e7c51a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 997.065580] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd549bc0-c83f-4aed-aebe-6aa5208aa1b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.087589] env[62521]: DEBUG oslo_vmware.api [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 997.087589] env[62521]: value = "task-1318952" [ 997.087589] env[62521]: _type = "Task" [ 997.087589] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.098950] env[62521]: DEBUG oslo_vmware.api [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318952, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.104235] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e83fe-3a7f-2c5a-2856-75f5c36883c3, 'name': SearchDatastore_Task, 'duration_secs': 0.016914} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.105051] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee27761c-4ec8-47eb-873c-136ab1447a7f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.110176] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 997.110176] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527f3710-4db5-aa58-404e-e75dd622dfcc" [ 997.110176] env[62521]: _type = "Task" [ 997.110176] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.118154] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527f3710-4db5-aa58-404e-e75dd622dfcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.253420] env[62521]: DEBUG oslo_concurrency.lockutils [req-b89f044e-9b70-41ac-a593-7925997f714e req-c029082b-bb9e-45df-9004-0c802683b52b service nova] Releasing lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.338177] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318951, 'name': ReconfigVM_Task, 'duration_secs': 0.390645} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.338520] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 7cdeed76-e5b3-4959-b6a6-610d23233e1b/7cdeed76-e5b3-4959-b6a6-610d23233e1b.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.339147] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7047947-ceb4-4c22-9450-d2f7f1818670 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.346170] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 997.346170] env[62521]: value = "task-1318953" [ 997.346170] env[62521]: _type = "Task" [ 997.346170] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.354276] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318953, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.362020] env[62521]: DEBUG nova.network.neutron [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Successfully created port: b0a026a3-abe7-4087-be49-669d7f6ba99e {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.497097] env[62521]: DEBUG nova.compute.manager [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 997.499833] env[62521]: DEBUG nova.network.neutron [-] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.592761] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.592910] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance d574f077-b221-4f91-8b54-0915421cb36f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.593042] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 11e61e72-a311-4c43-bf53-df82ad9d70f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.593173] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.593290] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance b8158051-98b6-47d8-84c5-d1d56515afe8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.593403] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.593564] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.593701] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance a68d4cb7-bfad-48f6-8aeb-bca04248a0d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.593818] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance a0646a00-1f1c-4892-bb32-6212c90e9e95 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.593931] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 7cdeed76-e5b3-4959-b6a6-610d23233e1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.594059] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 69c5ff30-259d-4067-b5e2-ab727009d8ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.594172] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 31f926a8-777d-4882-afd6-1e4dba3f4e11 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.594279] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance aa266be6-2bf1-445f-a968-2c144d30ecfa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 997.594554] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 997.594606] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3008MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 997.602234] env[62521]: DEBUG oslo_vmware.api [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318952, 'name': ReconfigVM_Task, 'duration_secs': 0.240661} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.602516] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Reconfigured VM instance instance-00000042 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 997.608164] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a78aa0f8-9f14-4691-8e4c-4ca9ca5ccb19 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.629754] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527f3710-4db5-aa58-404e-e75dd622dfcc, 'name': SearchDatastore_Task, 'duration_secs': 0.013097} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.631119] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.631406] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 69c5ff30-259d-4067-b5e2-ab727009d8ad/69c5ff30-259d-4067-b5e2-ab727009d8ad.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 997.631731] env[62521]: DEBUG oslo_vmware.api [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 997.631731] env[62521]: value = "task-1318954" [ 997.631731] env[62521]: _type = "Task" [ 997.631731] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.634270] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.634512] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.634705] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f29db7f-0f37-4ac8-b72a-f71e071f5140 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.637087] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b708619-1c6a-4554-a192-702dd8ea09eb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.647795] env[62521]: DEBUG oslo_vmware.api [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318954, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.651478] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 997.651478] env[62521]: value = "task-1318955" [ 997.651478] env[62521]: _type = "Task" [ 997.651478] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.652796] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 997.652977] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 997.656543] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1bb1fa6-24e5-4672-97e1-efea7223a5b3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.663863] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.669027] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 997.669027] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529fa7f6-f502-b804-f641-3c77c6a0d74f" [ 997.669027] env[62521]: _type = "Task" [ 997.669027] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.675265] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529fa7f6-f502-b804-f641-3c77c6a0d74f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.792598] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b761ea5-d92b-43fe-8732-43046e4bcd21 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.802523] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06ea479-6eb3-41fc-a29a-990e2b0f6c65 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.840832] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9e6885-f39f-4d16-b210-57bfbf894c52 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.854221] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701f06f7-0b5c-465e-b02e-6dddf58fbb52 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.863770] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318953, 'name': Rename_Task, 'duration_secs': 0.156324} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.875137] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 997.875644] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 997.877099] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-37d1e7a5-2230-4878-9f5b-f52542a9796c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.884301] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 997.884301] env[62521]: value = "task-1318956" [ 997.884301] env[62521]: _type = "Task" [ 997.884301] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.892897] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318956, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.980015] env[62521]: DEBUG oslo_vmware.rw_handles [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52725636-997e-846b-20ed-1c47dd190b15/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 997.980995] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a216d7-6747-41ff-92c7-9dabbc945910 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.987269] env[62521]: DEBUG oslo_vmware.rw_handles [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52725636-997e-846b-20ed-1c47dd190b15/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 997.987434] env[62521]: ERROR oslo_vmware.rw_handles [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52725636-997e-846b-20ed-1c47dd190b15/disk-0.vmdk due to incomplete transfer. [ 997.987648] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e825f0ea-4827-4080-a819-3d27f88532ef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.994254] env[62521]: DEBUG oslo_vmware.rw_handles [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52725636-997e-846b-20ed-1c47dd190b15/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 997.994460] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Uploaded image 2d8e21ca-6710-453b-83a9-3f2485192c5d to the Glance image server {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 997.996911] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 997.997174] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5e162f4b-09ad-49ba-a00d-c8ad1ae80c07 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.002805] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 998.002805] env[62521]: value = "task-1318957" [ 998.002805] env[62521]: _type = "Task" [ 998.002805] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.010049] env[62521]: INFO nova.compute.manager [-] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Took 1.33 seconds to deallocate network for instance. [ 998.017467] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318957, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.147427] env[62521]: DEBUG oslo_vmware.api [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318954, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.161595] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318955, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.178867] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529fa7f6-f502-b804-f641-3c77c6a0d74f, 'name': SearchDatastore_Task, 'duration_secs': 0.013492} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.180146] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc008276-e0a6-460e-962b-204784b67569 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.186373] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 998.186373] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b0c169-1631-6bf9-1e6a-de7839e8732c" [ 998.186373] env[62521]: _type = "Task" [ 998.186373] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.195473] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b0c169-1631-6bf9-1e6a-de7839e8732c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.207586] env[62521]: DEBUG nova.compute.manager [req-803e6020-80bc-49c3-90b6-d75a63f9dc58 req-ba65f9a3-1ae8-4ff0-bf0f-cdd4f4330889 service nova] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Received event network-vif-deleted-ff67aa5f-1a3c-478f-bf15-cf0577756a05 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.395315] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318956, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.398819] env[62521]: ERROR nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [req-431842eb-5738-4705-99cf-c93200d38d3c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-431842eb-5738-4705-99cf-c93200d38d3c"}]} [ 998.415334] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 998.433224] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 998.433224] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 998.445474] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 998.465016] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 998.514721] env[62521]: DEBUG nova.compute.manager [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 998.516653] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318957, 'name': Destroy_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.523239] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.550200] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 998.550582] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 998.550769] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.550968] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 998.551194] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.551359] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 998.551579] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 998.551746] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 998.551920] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 998.552146] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 998.552366] env[62521]: DEBUG nova.virt.hardware [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.553273] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ea8159-8436-46ef-9a14-13ca126175e3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.564365] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3542598-9b83-46b2-b8e2-b1beb16d0888 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.664593] env[62521]: DEBUG oslo_vmware.api [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318954, 'name': ReconfigVM_Task, 'duration_secs': 1.027946} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.664593] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282136', 'volume_id': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'name': 'volume-a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '73bc1cf1-7ee1-4a9b-8214-2dc752b700ad', 'attached_at': '', 'detached_at': '', 'volume_id': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c', 'serial': 'a3b15dfc-0fe5-4fef-ae6a-cb6f06553d4c'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 998.672197] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318955, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.783644} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.672662] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 69c5ff30-259d-4067-b5e2-ab727009d8ad/69c5ff30-259d-4067-b5e2-ab727009d8ad.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 998.678646] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.678646] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9151cabc-2d6a-4f78-813f-0bbe58ea6133 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.679510] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 998.679510] env[62521]: value = "task-1318958" [ 998.679510] env[62521]: _type = "Task" [ 998.679510] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.688602] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318958, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.698617] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b0c169-1631-6bf9-1e6a-de7839e8732c, 'name': SearchDatastore_Task, 'duration_secs': 0.061264} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.698617] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.698617] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 31f926a8-777d-4882-afd6-1e4dba3f4e11/31f926a8-777d-4882-afd6-1e4dba3f4e11.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 998.698617] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3fb8620d-c106-4ffb-b595-0a645413608a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.707333] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 998.707333] env[62521]: value = "task-1318959" [ 998.707333] env[62521]: _type = "Task" [ 998.707333] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.722123] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318959, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.747135] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa6f86b-3088-45ce-a7cf-793d1e3fb8d5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.757772] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1c519b-5003-4bac-bb67-48592a2028f5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.808879] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8188781b-ac7b-46a2-a3c4-2e7f4126f1a2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.820791] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b33537-e329-4742-9d90-2eb9dce6065a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.843905] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 998.894535] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318956, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.957339] env[62521]: DEBUG nova.network.neutron [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Successfully updated port: b0a026a3-abe7-4087-be49-669d7f6ba99e {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.015639] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318957, 'name': Destroy_Task, 'duration_secs': 0.514982} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.015908] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Destroyed the VM [ 999.016206] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 999.018067] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-da4b7060-da7b-44f7-a4c8-f3b814695d78 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.023633] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 999.023633] env[62521]: value = "task-1318960" [ 999.023633] env[62521]: _type = "Task" [ 999.023633] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.032200] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318960, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.190993] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318958, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079607} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.191324] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.192173] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf49a138-de1b-4a34-b344-f9fb521bc57f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.215405] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 69c5ff30-259d-4067-b5e2-ab727009d8ad/69c5ff30-259d-4067-b5e2-ab727009d8ad.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.215833] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89c37d3b-ef01-421e-afdc-3e1bedbb6ef6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.241595] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318959, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.243168] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 999.243168] env[62521]: value = "task-1318961" [ 999.243168] env[62521]: _type = "Task" [ 999.243168] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.252417] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318961, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.265866] env[62521]: DEBUG nova.objects.instance [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'flavor' on Instance uuid 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.382488] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 120 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 999.382488] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 120 to 121 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 999.382488] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 999.396376] env[62521]: DEBUG oslo_vmware.api [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318956, 'name': PowerOnVM_Task, 'duration_secs': 1.123976} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.396783] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.396977] env[62521]: INFO nova.compute.manager [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Took 9.92 seconds to spawn the instance on the hypervisor. [ 999.397128] env[62521]: DEBUG nova.compute.manager [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.398134] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce21eea-e3b2-48d8-8ceb-014417ee4983 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.460387] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-aa266be6-2bf1-445f-a968-2c144d30ecfa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.460605] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-aa266be6-2bf1-445f-a968-2c144d30ecfa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.460724] env[62521]: DEBUG nova.network.neutron [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 999.534420] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318960, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.726010] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318959, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.810008} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.726276] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 31f926a8-777d-4882-afd6-1e4dba3f4e11/31f926a8-777d-4882-afd6-1e4dba3f4e11.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.726459] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.726778] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-579b8060-e566-4de7-9af6-d74c0e02ab42 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.734863] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 999.734863] env[62521]: value = "task-1318962" [ 999.734863] env[62521]: _type = "Task" [ 999.734863] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.741884] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318962, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.751876] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318961, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.891387] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 999.891687] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.354s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.891872] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.369s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.892119] env[62521]: DEBUG nova.objects.instance [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lazy-loading 'resources' on Instance uuid a0646a00-1f1c-4892-bb32-6212c90e9e95 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.915043] env[62521]: INFO nova.compute.manager [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Took 21.13 seconds to build instance. [ 999.993349] env[62521]: DEBUG nova.network.neutron [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1000.036819] env[62521]: DEBUG oslo_vmware.api [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318960, 'name': RemoveSnapshot_Task, 'duration_secs': 0.67171} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.036931] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1000.037212] env[62521]: INFO nova.compute.manager [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Took 15.04 seconds to snapshot the instance on the hypervisor. [ 1000.129750] env[62521]: DEBUG nova.network.neutron [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Updating instance_info_cache with network_info: [{"id": "b0a026a3-abe7-4087-be49-669d7f6ba99e", "address": "fa:16:3e:6e:be:31", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0a026a3-ab", "ovs_interfaceid": "b0a026a3-abe7-4087-be49-669d7f6ba99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.245038] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318962, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068464} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.248107] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1000.249300] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703a1b9e-bfee-47d3-9c4f-922d9ca9a795 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.256549] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318961, 'name': ReconfigVM_Task, 'duration_secs': 0.523998} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.266340] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 69c5ff30-259d-4067-b5e2-ab727009d8ad/69c5ff30-259d-4067-b5e2-ab727009d8ad.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.277107] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 31f926a8-777d-4882-afd6-1e4dba3f4e11/31f926a8-777d-4882-afd6-1e4dba3f4e11.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.277399] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e8ea41d-08c9-47d0-8631-a284bfb52c7e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.282173] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9f98ab04-5b57-4959-b384-3c0f775e9984 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.875s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.282958] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f95b2d9c-7354-4261-a25c-f513e92473f5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.300907] env[62521]: DEBUG nova.compute.manager [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Received event network-vif-plugged-b0a026a3-abe7-4087-be49-669d7f6ba99e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.300907] env[62521]: DEBUG oslo_concurrency.lockutils [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] Acquiring lock "aa266be6-2bf1-445f-a968-2c144d30ecfa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.300907] env[62521]: DEBUG oslo_concurrency.lockutils [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] Lock "aa266be6-2bf1-445f-a968-2c144d30ecfa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.301292] env[62521]: DEBUG oslo_concurrency.lockutils [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] Lock "aa266be6-2bf1-445f-a968-2c144d30ecfa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.301292] env[62521]: DEBUG nova.compute.manager [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] No waiting events found dispatching network-vif-plugged-b0a026a3-abe7-4087-be49-669d7f6ba99e {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1000.301381] env[62521]: WARNING nova.compute.manager [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Received unexpected event network-vif-plugged-b0a026a3-abe7-4087-be49-669d7f6ba99e for instance with vm_state building and task_state spawning. [ 1000.301541] env[62521]: DEBUG nova.compute.manager [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Received event network-changed-b0a026a3-abe7-4087-be49-669d7f6ba99e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.301707] env[62521]: DEBUG nova.compute.manager [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Refreshing instance network info cache due to event network-changed-b0a026a3-abe7-4087-be49-669d7f6ba99e. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1000.301884] env[62521]: DEBUG oslo_concurrency.lockutils [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] Acquiring lock "refresh_cache-aa266be6-2bf1-445f-a968-2c144d30ecfa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.307765] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1000.307765] env[62521]: value = "task-1318963" [ 1000.307765] env[62521]: _type = "Task" [ 1000.307765] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.309229] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1000.309229] env[62521]: value = "task-1318964" [ 1000.309229] env[62521]: _type = "Task" [ 1000.309229] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.320495] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318963, 'name': Rename_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.323722] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318964, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.419089] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e8800ad-b35a-41a3-964f-4509000fe2e5 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.636s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.595326] env[62521]: DEBUG nova.compute.manager [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Found 3 images (rotation: 2) {{(pid=62521) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1000.595555] env[62521]: DEBUG nova.compute.manager [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Rotating out 1 backups {{(pid=62521) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1000.595724] env[62521]: DEBUG nova.compute.manager [None req-d77d68f5-bc2b-467d-8123-009a9a927e85 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Deleting image 07530e24-0ad3-44e9-9fb4-ee46d23c744c {{(pid=62521) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1000.625718] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccb0dc6-00b6-46b7-a931-ca0c625f1817 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.635089] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-aa266be6-2bf1-445f-a968-2c144d30ecfa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.635089] env[62521]: DEBUG nova.compute.manager [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Instance network_info: |[{"id": "b0a026a3-abe7-4087-be49-669d7f6ba99e", "address": "fa:16:3e:6e:be:31", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0a026a3-ab", "ovs_interfaceid": "b0a026a3-abe7-4087-be49-669d7f6ba99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1000.637295] env[62521]: DEBUG oslo_concurrency.lockutils [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] Acquired lock "refresh_cache-aa266be6-2bf1-445f-a968-2c144d30ecfa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.637695] env[62521]: DEBUG nova.network.neutron [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Refreshing network info cache for port b0a026a3-abe7-4087-be49-669d7f6ba99e {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1000.640096] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:be:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0a026a3-abe7-4087-be49-669d7f6ba99e', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1000.658251] env[62521]: DEBUG oslo.service.loopingcall [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.663016] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1000.664689] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba000206-530c-4aac-b63e-f88fb61c0691 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.673061] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d7bbeaf5-f044-49bb-bad2-8a5dc28a80ac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.727626] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8662e756-eb6b-4bc4-83a4-2120e09513eb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.730455] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1000.730455] env[62521]: value = "task-1318965" [ 1000.730455] env[62521]: _type = "Task" [ 1000.730455] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.739281] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3563e34-0432-4eca-8303-64a5d86a05a6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.746204] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318965, 'name': CreateVM_Task} progress is 15%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.755978] env[62521]: DEBUG nova.compute.provider_tree [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.798786] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.799103] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.799371] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.799606] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.799830] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.802132] env[62521]: INFO nova.compute.manager [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Terminating instance [ 1000.804213] env[62521]: DEBUG nova.compute.manager [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.804435] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.805400] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e1580c-416b-49d8-b135-6cb243b05e31 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.815754] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.819265] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20a727f7-2d0c-4133-9987-578fdca0b549 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.823658] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318963, 'name': Rename_Task, 'duration_secs': 0.14493} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.824532] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1000.824786] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e7438e9-5ed1-480a-a5d9-ad188acefd88 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.829312] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318964, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.831246] env[62521]: DEBUG oslo_vmware.api [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1000.831246] env[62521]: value = "task-1318966" [ 1000.831246] env[62521]: _type = "Task" [ 1000.831246] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.835855] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1000.835855] env[62521]: value = "task-1318967" [ 1000.835855] env[62521]: _type = "Task" [ 1000.835855] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.844745] env[62521]: DEBUG oslo_vmware.api [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318966, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.850095] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318967, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.923556] env[62521]: DEBUG nova.network.neutron [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Updated VIF entry in instance network info cache for port b0a026a3-abe7-4087-be49-669d7f6ba99e. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1000.923948] env[62521]: DEBUG nova.network.neutron [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Updating instance_info_cache with network_info: [{"id": "b0a026a3-abe7-4087-be49-669d7f6ba99e", "address": "fa:16:3e:6e:be:31", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0a026a3-ab", "ovs_interfaceid": "b0a026a3-abe7-4087-be49-669d7f6ba99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.240411] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318965, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.258581] env[62521]: DEBUG nova.scheduler.client.report [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.324969] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318964, 'name': ReconfigVM_Task, 'duration_secs': 0.765455} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.325320] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 31f926a8-777d-4882-afd6-1e4dba3f4e11/31f926a8-777d-4882-afd6-1e4dba3f4e11.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.325954] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a108d66-8904-4c7c-aef6-c5940b54c4f6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.332048] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1001.332048] env[62521]: value = "task-1318968" [ 1001.332048] env[62521]: _type = "Task" [ 1001.332048] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.344480] env[62521]: DEBUG oslo_vmware.api [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318966, 'name': PowerOffVM_Task, 'duration_secs': 0.316675} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.348122] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.348347] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.348638] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318968, 'name': Rename_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.349215] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de935981-507a-4851-90fb-a1845f59b227 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.353502] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318967, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.427497] env[62521]: DEBUG oslo_concurrency.lockutils [req-ee9f1a68-3050-4c61-a07e-2f7185daeef2 req-f9007011-893b-4965-b372-022a2de16029 service nova] Releasing lock "refresh_cache-aa266be6-2bf1-445f-a968-2c144d30ecfa" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.470828] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.471156] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.471440] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleting the datastore file [datastore1] 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.471755] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b09d0761-b743-47ef-865c-3791946cada1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.479985] env[62521]: DEBUG oslo_vmware.api [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1001.479985] env[62521]: value = "task-1318970" [ 1001.479985] env[62521]: _type = "Task" [ 1001.479985] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.489367] env[62521]: DEBUG oslo_vmware.api [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318970, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.645688] env[62521]: DEBUG oslo_concurrency.lockutils [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.645938] env[62521]: DEBUG oslo_concurrency.lockutils [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.740050] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318965, 'name': CreateVM_Task, 'duration_secs': 0.656467} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.740245] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1001.740945] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.741138] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.741464] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1001.741721] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6b1e9ac-ef2f-4df7-9ef9-7bb226b19d3f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.745961] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1001.745961] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ea75d0-f616-47cd-c3f2-102ed6f865e2" [ 1001.745961] env[62521]: _type = "Task" [ 1001.745961] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.753311] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ea75d0-f616-47cd-c3f2-102ed6f865e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.764122] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.872s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.785687] env[62521]: INFO nova.scheduler.client.report [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Deleted allocations for instance a0646a00-1f1c-4892-bb32-6212c90e9e95 [ 1001.847267] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318968, 'name': Rename_Task, 'duration_secs': 0.16622} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.850532] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.850854] env[62521]: DEBUG oslo_vmware.api [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318967, 'name': PowerOnVM_Task, 'duration_secs': 0.643146} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.851116] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62e2e356-1fd2-4e9a-a34f-e74644da170e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.852783] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1001.853037] env[62521]: INFO nova.compute.manager [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Took 9.95 seconds to spawn the instance on the hypervisor. [ 1001.853265] env[62521]: DEBUG nova.compute.manager [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.854066] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98f8d88-5f7b-4c25-9b63-071db112d58f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.863603] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1001.863603] env[62521]: value = "task-1318971" [ 1001.863603] env[62521]: _type = "Task" [ 1001.863603] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.871222] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318971, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.990019] env[62521]: DEBUG oslo_vmware.api [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1318970, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234638} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.990421] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.990615] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.990792] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.990972] env[62521]: INFO nova.compute.manager [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1001.991241] env[62521]: DEBUG oslo.service.loopingcall [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.991442] env[62521]: DEBUG nova.compute.manager [-] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1001.991539] env[62521]: DEBUG nova.network.neutron [-] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1002.149543] env[62521]: DEBUG nova.compute.utils [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1002.256596] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ea75d0-f616-47cd-c3f2-102ed6f865e2, 'name': SearchDatastore_Task, 'duration_secs': 0.013123} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.256912] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.257194] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.257437] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.257588] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.257772] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.258055] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b3fb67de-22bb-4af4-aa29-901ba33b7c9a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.275020] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.275303] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1002.276723] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63b6ee0d-c460-476d-95c4-3ccacc9d680f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.282618] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1002.282618] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52253b90-fdcb-59e7-65dd-d17ec38a66b2" [ 1002.282618] env[62521]: _type = "Task" [ 1002.282618] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.293632] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52253b90-fdcb-59e7-65dd-d17ec38a66b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.294313] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b781a7cf-750a-48ef-ab03-c6cab5022ea8 tempest-ServerDiskConfigTestJSON-282309071 tempest-ServerDiskConfigTestJSON-282309071-project-member] Lock "a0646a00-1f1c-4892-bb32-6212c90e9e95" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.743s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.373767] env[62521]: INFO nova.compute.manager [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Took 22.08 seconds to build instance. [ 1002.380621] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318971, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.653754] env[62521]: DEBUG oslo_concurrency.lockutils [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.764766] env[62521]: DEBUG nova.compute.manager [req-d09f0951-7158-49a0-874a-a178b22ef22a req-01c5a669-bc50-4ec8-921d-88dddf9fcba9 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Received event network-vif-deleted-72671c1a-eabd-4a80-a591-17b26cea0249 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.765019] env[62521]: INFO nova.compute.manager [req-d09f0951-7158-49a0-874a-a178b22ef22a req-01c5a669-bc50-4ec8-921d-88dddf9fcba9 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Neutron deleted interface 72671c1a-eabd-4a80-a591-17b26cea0249; detaching it from the instance and deleting it from the info cache [ 1002.766056] env[62521]: DEBUG nova.network.neutron [req-d09f0951-7158-49a0-874a-a178b22ef22a req-01c5a669-bc50-4ec8-921d-88dddf9fcba9 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.792921] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52253b90-fdcb-59e7-65dd-d17ec38a66b2, 'name': SearchDatastore_Task, 'duration_secs': 0.014747} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.793780] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-269ca09f-723f-4bc5-bc5a-ccb911cc87ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.799202] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1002.799202] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5204892d-2d06-ff8d-fddb-96e91579f085" [ 1002.799202] env[62521]: _type = "Task" [ 1002.799202] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.807946] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5204892d-2d06-ff8d-fddb-96e91579f085, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.878809] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bced1678-9be6-4dc6-803d-964310afae04 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "69c5ff30-259d-4067-b5e2-ab727009d8ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.601s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.879123] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318971, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.971135] env[62521]: DEBUG nova.network.neutron [-] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.268429] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aed408b3-c7fb-41ed-bd5f-04f60a39c9bf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.277752] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f3d768-e3ee-4c81-bf4c-2f3de6f23c09 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.317651] env[62521]: DEBUG nova.compute.manager [req-d09f0951-7158-49a0-874a-a178b22ef22a req-01c5a669-bc50-4ec8-921d-88dddf9fcba9 service nova] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Detach interface failed, port_id=72671c1a-eabd-4a80-a591-17b26cea0249, reason: Instance 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1003.322643] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5204892d-2d06-ff8d-fddb-96e91579f085, 'name': SearchDatastore_Task, 'duration_secs': 0.017465} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.322643] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.322643] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] aa266be6-2bf1-445f-a968-2c144d30ecfa/aa266be6-2bf1-445f-a968-2c144d30ecfa.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1003.322881] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d77e5c9e-13ab-43dd-990d-d59d21d7c8ec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.329249] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1003.329249] env[62521]: value = "task-1318972" [ 1003.329249] env[62521]: _type = "Task" [ 1003.329249] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.336939] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318972, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.386588] env[62521]: DEBUG oslo_vmware.api [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318971, 'name': PowerOnVM_Task, 'duration_secs': 1.120412} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.386891] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1003.387176] env[62521]: INFO nova.compute.manager [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Took 9.11 seconds to spawn the instance on the hypervisor. [ 1003.387391] env[62521]: DEBUG nova.compute.manager [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.388229] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c66a0b0-27bc-472c-83dc-f27f727697f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.428117] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "69c5ff30-259d-4067-b5e2-ab727009d8ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.428520] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "69c5ff30-259d-4067-b5e2-ab727009d8ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.428823] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "69c5ff30-259d-4067-b5e2-ab727009d8ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.429102] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "69c5ff30-259d-4067-b5e2-ab727009d8ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.429349] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "69c5ff30-259d-4067-b5e2-ab727009d8ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.431731] env[62521]: INFO nova.compute.manager [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Terminating instance [ 1003.433709] env[62521]: DEBUG nova.compute.manager [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.433903] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.434820] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525ffeee-b712-4def-b0a4-ce1eae665c8a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.443117] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.443355] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c73b617-80a0-49eb-ae3b-d60968c422fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.449071] env[62521]: DEBUG oslo_vmware.api [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1003.449071] env[62521]: value = "task-1318973" [ 1003.449071] env[62521]: _type = "Task" [ 1003.449071] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.456887] env[62521]: DEBUG oslo_vmware.api [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318973, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.478312] env[62521]: INFO nova.compute.manager [-] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Took 1.48 seconds to deallocate network for instance. [ 1003.840909] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318972, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.904367] env[62521]: INFO nova.compute.manager [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Took 21.49 seconds to build instance. [ 1003.958536] env[62521]: DEBUG oslo_vmware.api [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318973, 'name': PowerOffVM_Task, 'duration_secs': 0.4851} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.958810] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.958985] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.959257] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06d10bf1-23f7-4a35-8894-8f19548e5f26 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.981100] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.981361] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.981582] env[62521]: DEBUG nova.objects.instance [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'resources' on Instance uuid 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.998436] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.998680] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.998862] env[62521]: DEBUG nova.compute.manager [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.999532] env[62521]: DEBUG oslo_concurrency.lockutils [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.999741] env[62521]: DEBUG oslo_concurrency.lockutils [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.999961] env[62521]: INFO nova.compute.manager [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Attaching volume 510f1170-3625-496a-9477-442b0e639a96 to /dev/sdb [ 1004.003122] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18de3905-79dd-4166-9d74-32bcb6349aef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.013247] env[62521]: DEBUG nova.compute.manager [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62521) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1004.014222] env[62521]: DEBUG nova.objects.instance [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'flavor' on Instance uuid 11e61e72-a311-4c43-bf53-df82ad9d70f7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1004.033570] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c4223c-5b05-4195-9935-7dc2e35387c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.040857] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a390f3d8-2dfe-4c75-9c3f-ef71c562d56b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.054963] env[62521]: DEBUG nova.virt.block_device [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Updating existing volume attachment record: b88de0e0-1a1d-43af-8df8-254ff8fff306 {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1004.143619] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1004.143862] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1004.144133] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleting the datastore file [datastore2] 69c5ff30-259d-4067-b5e2-ab727009d8ad {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.144417] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db9e936b-5800-4613-b467-5e116e8930f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.151470] env[62521]: DEBUG oslo_vmware.api [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1004.151470] env[62521]: value = "task-1318975" [ 1004.151470] env[62521]: _type = "Task" [ 1004.151470] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.159824] env[62521]: DEBUG oslo_vmware.api [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318975, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.340729] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318972, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579915} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.340729] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] aa266be6-2bf1-445f-a968-2c144d30ecfa/aa266be6-2bf1-445f-a968-2c144d30ecfa.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1004.340729] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.340729] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85447e60-4010-4848-9bdc-a63d069a6b9e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.345449] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1004.345449] env[62521]: value = "task-1318979" [ 1004.345449] env[62521]: _type = "Task" [ 1004.345449] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.353032] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318979, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.406750] env[62521]: DEBUG oslo_concurrency.lockutils [None req-aa370551-c681-4b37-8a6f-d8db8ecf6b4a tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lock "31f926a8-777d-4882-afd6-1e4dba3f4e11" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.995s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.523175] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1004.523493] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56c2eb16-2625-4170-834e-11d3ea408de2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.538759] env[62521]: DEBUG oslo_vmware.api [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1004.538759] env[62521]: value = "task-1318980" [ 1004.538759] env[62521]: _type = "Task" [ 1004.538759] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.549566] env[62521]: DEBUG oslo_vmware.api [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318980, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.662445] env[62521]: DEBUG oslo_vmware.api [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318975, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169515} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.662718] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.662905] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.666208] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.666461] env[62521]: INFO nova.compute.manager [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Took 1.23 seconds to destroy the instance on the hypervisor. [ 1004.666723] env[62521]: DEBUG oslo.service.loopingcall [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.666927] env[62521]: DEBUG nova.compute.manager [-] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.667057] env[62521]: DEBUG nova.network.neutron [-] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1004.676041] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05ea6aa-34bc-42e7-925d-5023c07f34b8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.683563] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4aa53d-3c38-4130-a52e-f7b429ab1083 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.716880] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60521a91-f184-4fbc-a3dd-dfb4308ddfa7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.725500] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9841d086-a3c3-41e0-9ca3-6d139b52d89c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.741502] env[62521]: DEBUG nova.compute.provider_tree [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.855260] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318979, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066916} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.855578] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.856419] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d458a6-ec7a-4664-a4f0-8b625c01ddcb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.877519] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] aa266be6-2bf1-445f-a968-2c144d30ecfa/aa266be6-2bf1-445f-a968-2c144d30ecfa.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.877744] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-361b58e0-8ac5-4814-916f-118118e88a2c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.897337] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1004.897337] env[62521]: value = "task-1318981" [ 1004.897337] env[62521]: _type = "Task" [ 1004.897337] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.904708] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318981, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.969776] env[62521]: INFO nova.compute.manager [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Rescuing [ 1004.970154] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquiring lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.970418] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquired lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.970526] env[62521]: DEBUG nova.network.neutron [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1005.050234] env[62521]: DEBUG oslo_vmware.api [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318980, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.095323] env[62521]: DEBUG nova.compute.manager [req-313f65b8-233d-453d-a4f9-f5bef81b9df7 req-552c4545-5d61-4823-870e-c28db61f5aa9 service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Received event network-vif-deleted-b4b87574-5fdf-40a2-b2ac-a97129208f8a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.095514] env[62521]: INFO nova.compute.manager [req-313f65b8-233d-453d-a4f9-f5bef81b9df7 req-552c4545-5d61-4823-870e-c28db61f5aa9 service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Neutron deleted interface b4b87574-5fdf-40a2-b2ac-a97129208f8a; detaching it from the instance and deleting it from the info cache [ 1005.095692] env[62521]: DEBUG nova.network.neutron [req-313f65b8-233d-453d-a4f9-f5bef81b9df7 req-552c4545-5d61-4823-870e-c28db61f5aa9 service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.918325] env[62521]: DEBUG nova.scheduler.client.report [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.923329] env[62521]: DEBUG nova.network.neutron [-] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.930412] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5aa1291e-5aa6-4113-988f-71da0f30e28a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.939173] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318981, 'name': ReconfigVM_Task, 'duration_secs': 0.326184} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.942854] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Reconfigured VM instance instance-0000005d to attach disk [datastore2] aa266be6-2bf1-445f-a968-2c144d30ecfa/aa266be6-2bf1-445f-a968-2c144d30ecfa.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.943609] env[62521]: DEBUG oslo_vmware.api [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1318980, 'name': PowerOffVM_Task, 'duration_secs': 0.544244} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.944074] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1da2526d-0547-4be2-97d4-28bf99ff09c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.948095] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7d2020-c503-4822-a8a5-e3efce80a909 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.958237] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1005.958441] env[62521]: DEBUG nova.compute.manager [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.961787] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd7d5e4-97b6-46b1-848b-0cde951822d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.970986] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1005.970986] env[62521]: value = "task-1318982" [ 1005.970986] env[62521]: _type = "Task" [ 1005.970986] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.983360] env[62521]: DEBUG nova.compute.manager [req-313f65b8-233d-453d-a4f9-f5bef81b9df7 req-552c4545-5d61-4823-870e-c28db61f5aa9 service nova] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Detach interface failed, port_id=b4b87574-5fdf-40a2-b2ac-a97129208f8a, reason: Instance 69c5ff30-259d-4067-b5e2-ab727009d8ad could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1005.988757] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318982, 'name': Rename_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.130072] env[62521]: DEBUG nova.network.neutron [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updating instance_info_cache with network_info: [{"id": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "address": "fa:16:3e:82:b0:68", "network": {"id": "e84e7dd3-8f72-4865-86ba-63a68f5178cf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1633868122-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f513dc5025974607a90e67621154bccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75ffddaa-cb", "ovs_interfaceid": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.427686] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.444s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.428105] env[62521]: INFO nova.compute.manager [-] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Took 1.76 seconds to deallocate network for instance. [ 1006.450137] env[62521]: INFO nova.scheduler.client.report [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleted allocations for instance 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad [ 1006.478364] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ebf030a1-3dfb-49cf-afb9-48516375cdb9 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.480s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.488218] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318982, 'name': Rename_Task, 'duration_secs': 0.359045} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.488512] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1006.488757] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f5764c7-0765-4b9d-a601-06d82cf79f11 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.495118] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1006.495118] env[62521]: value = "task-1318984" [ 1006.495118] env[62521]: _type = "Task" [ 1006.495118] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.504405] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318984, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.632618] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Releasing lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.940184] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.940864] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.941018] env[62521]: DEBUG nova.objects.instance [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lazy-loading 'resources' on Instance uuid 69c5ff30-259d-4067-b5e2-ab727009d8ad {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.958074] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f83a501-f4d8-4114-9932-fdff4c71a998 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "73bc1cf1-7ee1-4a9b-8214-2dc752b700ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.159s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.008474] env[62521]: DEBUG oslo_vmware.api [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318984, 'name': PowerOnVM_Task, 'duration_secs': 0.47039} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.008869] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1007.009176] env[62521]: INFO nova.compute.manager [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Took 8.49 seconds to spawn the instance on the hypervisor. [ 1007.009429] env[62521]: DEBUG nova.compute.manager [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.010459] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5024c61-de48-4df2-af63-4e966be0f984 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.166663] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.167172] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30d5452e-6a7b-4315-94ea-369b9d42554d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.174437] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1007.174437] env[62521]: value = "task-1318985" [ 1007.174437] env[62521]: _type = "Task" [ 1007.174437] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.184532] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318985, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.531510] env[62521]: INFO nova.compute.manager [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Took 23.02 seconds to build instance. [ 1007.608432] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b03fa01-8139-4807-824a-cfe7ffdf8bcc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.616768] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a147769-d29d-488b-ab1e-32a1e6cc902a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.648477] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e60e16-156c-4542-a53a-93a143ac7551 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.656261] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89812c6-b4cd-4d99-9a87-968bc9435fbe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.671878] env[62521]: DEBUG nova.compute.provider_tree [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1007.683283] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318985, 'name': PowerOffVM_Task, 'duration_secs': 0.220058} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.683548] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.684312] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda81a76-6996-46d1-af9a-f4afa821d02a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.706320] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f4ed65-3637-4952-a3f2-e18d58f3bf51 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.743925] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.744252] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7a495f8-1beb-4f6e-8a9f-7691efaf2102 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.751885] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1007.751885] env[62521]: value = "task-1318986" [ 1007.751885] env[62521]: _type = "Task" [ 1007.751885] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.759466] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318986, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.033582] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1c83d44-a853-431e-aec1-d3c95c0d9fe0 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "aa266be6-2bf1-445f-a968-2c144d30ecfa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.528s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.195977] env[62521]: ERROR nova.scheduler.client.report [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [req-5b67f9d1-7182-4e76-b126-094e3647d4a5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5b67f9d1-7182-4e76-b126-094e3647d4a5"}]} [ 1008.214636] env[62521]: DEBUG nova.scheduler.client.report [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1008.238659] env[62521]: DEBUG nova.scheduler.client.report [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1008.238936] env[62521]: DEBUG nova.compute.provider_tree [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1008.258770] env[62521]: DEBUG nova.scheduler.client.report [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1008.265208] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1008.265208] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.265208] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.265759] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.265759] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.265909] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f527d428-8a6c-4b95-9dae-f78a2af8e08f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.277603] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.277823] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1008.278618] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a3c7f9b-8d82-49d8-95a7-799d83012325 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.286368] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1008.286368] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52226e45-f0e7-850c-6e58-7f4b06b8d8c0" [ 1008.286368] env[62521]: _type = "Task" [ 1008.286368] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.289031] env[62521]: DEBUG nova.scheduler.client.report [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1008.294559] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52226e45-f0e7-850c-6e58-7f4b06b8d8c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.402036] env[62521]: DEBUG nova.compute.manager [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Stashing vm_state: stopped {{(pid=62521) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1008.521194] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea03c67-fceb-41d3-b9a8-86183af69b12 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.529226] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba596662-6d4a-44bc-9dde-2a8879e49f08 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.563751] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1242fdaa-f721-406c-b8b0-83c9557fa25a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.572087] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7baee468-b7a1-4dff-86a5-f42338acc390 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.587064] env[62521]: DEBUG nova.compute.provider_tree [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1008.600826] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1008.601071] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282168', 'volume_id': '510f1170-3625-496a-9477-442b0e639a96', 'name': 'volume-510f1170-3625-496a-9477-442b0e639a96', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7cdeed76-e5b3-4959-b6a6-610d23233e1b', 'attached_at': '', 'detached_at': '', 'volume_id': '510f1170-3625-496a-9477-442b0e639a96', 'serial': '510f1170-3625-496a-9477-442b0e639a96'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1008.602129] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e1417f-e8f9-4132-aa99-2477f68fcc66 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.622281] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790acc89-eaaa-4635-a621-edda4d09fc34 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.646833] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] volume-510f1170-3625-496a-9477-442b0e639a96/volume-510f1170-3625-496a-9477-442b0e639a96.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1008.646833] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6df80c21-29e8-4b20-b867-115fcf35d04c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.665052] env[62521]: DEBUG oslo_vmware.api [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1008.665052] env[62521]: value = "task-1318987" [ 1008.665052] env[62521]: _type = "Task" [ 1008.665052] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.674095] env[62521]: DEBUG oslo_vmware.api [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.807396] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52226e45-f0e7-850c-6e58-7f4b06b8d8c0, 'name': SearchDatastore_Task, 'duration_secs': 0.026329} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.808309] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d66d337-ece0-4125-ad08-efcb5e598fe3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.819021] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1008.819021] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d40bc9-948b-5180-2b71-7c2680028882" [ 1008.819021] env[62521]: _type = "Task" [ 1008.819021] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.825757] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d40bc9-948b-5180-2b71-7c2680028882, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.928866] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.122532] env[62521]: DEBUG nova.scheduler.client.report [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 123 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1009.122893] env[62521]: DEBUG nova.compute.provider_tree [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 123 to 124 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1009.122980] env[62521]: DEBUG nova.compute.provider_tree [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.175463] env[62521]: DEBUG oslo_vmware.api [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318987, 'name': ReconfigVM_Task, 'duration_secs': 0.443248} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.175765] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Reconfigured VM instance instance-0000005a to attach disk [datastore1] volume-510f1170-3625-496a-9477-442b0e639a96/volume-510f1170-3625-496a-9477-442b0e639a96.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1009.182199] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d29454cb-1984-4560-838a-93d90eeb7f2d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.202230] env[62521]: DEBUG oslo_vmware.api [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1009.202230] env[62521]: value = "task-1318988" [ 1009.202230] env[62521]: _type = "Task" [ 1009.202230] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.212295] env[62521]: DEBUG oslo_vmware.api [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318988, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.301595] env[62521]: INFO nova.compute.manager [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Rebuilding instance [ 1009.330661] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d40bc9-948b-5180-2b71-7c2680028882, 'name': SearchDatastore_Task, 'duration_secs': 0.011495} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.335013] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.335319] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 31f926a8-777d-4882-afd6-1e4dba3f4e11/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk. {{(pid=62521) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1009.335988] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b4d9089-cba2-4b2c-936a-642721ca9c1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.343064] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1009.343064] env[62521]: value = "task-1318989" [ 1009.343064] env[62521]: _type = "Task" [ 1009.343064] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.352304] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318989, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.361550] env[62521]: DEBUG nova.compute.manager [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.361955] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac2708a-7858-4dad-a04f-6f43759da306 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.628663] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.688s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.635948] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.702s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.664147] env[62521]: INFO nova.scheduler.client.report [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted allocations for instance 69c5ff30-259d-4067-b5e2-ab727009d8ad [ 1009.713371] env[62521]: DEBUG oslo_vmware.api [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318988, 'name': ReconfigVM_Task, 'duration_secs': 0.164563} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.714535] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282168', 'volume_id': '510f1170-3625-496a-9477-442b0e639a96', 'name': 'volume-510f1170-3625-496a-9477-442b0e639a96', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7cdeed76-e5b3-4959-b6a6-610d23233e1b', 'attached_at': '', 'detached_at': '', 'volume_id': '510f1170-3625-496a-9477-442b0e639a96', 'serial': '510f1170-3625-496a-9477-442b0e639a96'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1009.855943] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318989, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.873727] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.873727] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb9872a0-aed5-4e11-a699-df49fe252fee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.879935] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1009.879935] env[62521]: value = "task-1318990" [ 1009.879935] env[62521]: _type = "Task" [ 1009.879935] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.895103] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318990, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.140969] env[62521]: INFO nova.compute.claims [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.173404] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a46135a9-8fff-4684-9cbf-b0a40a54a49e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "69c5ff30-259d-4067-b5e2-ab727009d8ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.745s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.354925] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318989, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.768856} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.355297] env[62521]: INFO nova.virt.vmwareapi.ds_util [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 31f926a8-777d-4882-afd6-1e4dba3f4e11/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk. [ 1010.356084] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb023e95-1cd1-436e-b5c2-35a301e79e54 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.383451] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 31f926a8-777d-4882-afd6-1e4dba3f4e11/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1010.383451] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8220e605-42c0-4ab8-9ab0-137a72790b3f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.404684] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318990, 'name': PowerOffVM_Task, 'duration_secs': 0.36001} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.411020] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.411020] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1010.411020] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1010.411020] env[62521]: value = "task-1318991" [ 1010.411020] env[62521]: _type = "Task" [ 1010.411020] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.411020] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b4ddd8-43b2-4004-b742-5a7125ec770b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.421256] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318991, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.423427] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.423675] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d10f99e1-2b3f-4cd2-bc35-6ec7a69008a7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.483182] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.484032] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.484708] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquiring lock "83fe58c9-920b-422d-be08-e5d53bf551d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.484848] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lock "83fe58c9-920b-422d-be08-e5d53bf551d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.489803] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.490026] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.490748] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleting the datastore file [datastore2] aa266be6-2bf1-445f-a968-2c144d30ecfa {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.490748] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-595aabf0-4378-4723-b7f6-2342db0a80d4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.497438] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1010.497438] env[62521]: value = "task-1318993" [ 1010.497438] env[62521]: _type = "Task" [ 1010.497438] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.505980] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.653707] env[62521]: INFO nova.compute.resource_tracker [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating resource usage from migration 1db23eae-760d-44f6-9059-4145c6530e07 [ 1010.659876] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.660214] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.662469] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.662717] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.662918] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.665321] env[62521]: INFO nova.compute.manager [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Terminating instance [ 1010.667635] env[62521]: DEBUG nova.compute.manager [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.667908] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1010.668862] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe94349-3ce9-4659-8385-dc1ed70506dd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.676536] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1010.679168] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8af42bea-0a78-489a-8a9d-6b7034753a06 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.685109] env[62521]: DEBUG oslo_vmware.api [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1010.685109] env[62521]: value = "task-1318994" [ 1010.685109] env[62521]: _type = "Task" [ 1010.685109] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.692868] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "interface-b8158051-98b6-47d8-84c5-d1d56515afe8-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.693231] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-b8158051-98b6-47d8-84c5-d1d56515afe8-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.693447] env[62521]: DEBUG nova.objects.instance [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'flavor' on Instance uuid b8158051-98b6-47d8-84c5-d1d56515afe8 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.697607] env[62521]: DEBUG oslo_vmware.api [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.768545] env[62521]: DEBUG nova.objects.instance [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lazy-loading 'flavor' on Instance uuid 7cdeed76-e5b3-4959-b6a6-610d23233e1b {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.815697] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d16a13-2628-4f69-80f9-a277b7393d62 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.822945] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f5c9c1-1d71-48a7-8821-4ac98e283b37 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.851725] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b50aae-75a7-4f89-9ca6-5e6ddfccab4a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.858585] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8dda784-c266-4a39-a857-f7ef5c16977e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.871106] env[62521]: DEBUG nova.compute.provider_tree [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.921417] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318991, 'name': ReconfigVM_Task, 'duration_secs': 0.312903} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.921709] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 31f926a8-777d-4882-afd6-1e4dba3f4e11/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1010.922579] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2ba486-cb8f-4bb2-b3fd-385234c51545 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.950797] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e3a7085-8972-4072-a584-82a5081d4255 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.965494] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1010.965494] env[62521]: value = "task-1318995" [ 1010.965494] env[62521]: _type = "Task" [ 1010.965494] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.973365] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318995, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.987405] env[62521]: DEBUG nova.compute.manager [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1010.990095] env[62521]: DEBUG nova.compute.manager [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.011401] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1318993, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145644} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.011668] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.011872] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.012067] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.194803] env[62521]: DEBUG oslo_vmware.api [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.270366] env[62521]: DEBUG nova.objects.instance [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'pci_requests' on Instance uuid b8158051-98b6-47d8-84c5-d1d56515afe8 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.275876] env[62521]: DEBUG oslo_concurrency.lockutils [None req-98b39aa2-98bf-4ba2-aab4-6a52e7c92d1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.276s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.374238] env[62521]: DEBUG nova.scheduler.client.report [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.475321] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318995, 'name': ReconfigVM_Task, 'duration_secs': 0.151075} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.475621] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1011.475872] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3bbd879-155a-405c-aae2-417baec0b3d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.482066] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1011.482066] env[62521]: value = "task-1318996" [ 1011.482066] env[62521]: _type = "Task" [ 1011.482066] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.489322] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318996, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.504288] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.504534] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.504747] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.504937] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.505134] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.506992] env[62521]: INFO nova.compute.manager [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Terminating instance [ 1011.508894] env[62521]: DEBUG nova.compute.manager [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1011.509149] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.509365] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed96e5e6-58f9-47fe-b7d8-2f893f205952 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.512196] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.513117] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.515201] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1011.515201] env[62521]: value = "task-1318997" [ 1011.515201] env[62521]: _type = "Task" [ 1011.515201] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.526953] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.695650] env[62521]: DEBUG oslo_vmware.api [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.774794] env[62521]: DEBUG nova.objects.base [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1011.775118] env[62521]: DEBUG nova.network.neutron [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1011.816376] env[62521]: DEBUG nova.policy [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ce6cad0be5f455eab09b1d348268329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '961924204a584b61a4a85a761821910d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1011.879739] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.248s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.879960] env[62521]: INFO nova.compute.manager [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Migrating [ 1011.886583] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.374s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.888054] env[62521]: INFO nova.compute.claims [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1011.993118] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318996, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.027215] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318997, 'name': PowerOffVM_Task, 'duration_secs': 0.289028} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.027842] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1012.027957] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1012.028935] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282168', 'volume_id': '510f1170-3625-496a-9477-442b0e639a96', 'name': 'volume-510f1170-3625-496a-9477-442b0e639a96', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7cdeed76-e5b3-4959-b6a6-610d23233e1b', 'attached_at': '', 'detached_at': '', 'volume_id': '510f1170-3625-496a-9477-442b0e639a96', 'serial': '510f1170-3625-496a-9477-442b0e639a96'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1012.028935] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f2ddc0-8a36-4d9a-81c8-1c690bbeef72 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.052018] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a8675e4-2ad1-407f-8ec8-272e02601af2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.058845] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d86ccd-357f-4970-88bb-f008864f12ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.062907] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.063154] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.063323] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.063513] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.063664] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.063814] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.064039] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.064212] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.064385] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.064553] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.064728] env[62521]: DEBUG nova.virt.hardware [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.065467] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36065d87-f0d1-4390-a990-8c20c279ab25 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.087258] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e1a344-b829-4eb3-993b-45eacb3b1b1b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.091887] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ee9792-b9d9-4735-9102-320647e8a530 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.109448] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] The volume has not been displaced from its original location: [datastore1] volume-510f1170-3625-496a-9477-442b0e639a96/volume-510f1170-3625-496a-9477-442b0e639a96.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1012.114865] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Reconfiguring VM instance instance-0000005a to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1012.125112] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ffaf6a56-e91e-4c0a-a117-d81fd930f92a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.138367] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:be:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0a026a3-abe7-4087-be49-669d7f6ba99e', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1012.146592] env[62521]: DEBUG oslo.service.loopingcall [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.147675] env[62521]: DEBUG nova.network.neutron [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Successfully created port: f5f1a879-9bd4-43e6-815a-307ae9cff9e7 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.149493] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1012.150211] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-800236e1-cecb-46bd-a378-0434c25ae91e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.167741] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1012.167741] env[62521]: value = "task-1318998" [ 1012.167741] env[62521]: _type = "Task" [ 1012.167741] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.176142] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318998, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.177381] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1012.177381] env[62521]: value = "task-1318999" [ 1012.177381] env[62521]: _type = "Task" [ 1012.177381] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.185519] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318999, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.193844] env[62521]: DEBUG oslo_vmware.api [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.400852] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.402164] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.402164] env[62521]: DEBUG nova.network.neutron [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.493251] env[62521]: DEBUG oslo_vmware.api [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1318996, 'name': PowerOnVM_Task, 'duration_secs': 0.555033} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.493620] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1012.496394] env[62521]: DEBUG nova.compute.manager [None req-9d49b049-132a-4068-ae00-cf0eff02ab2d tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1012.497391] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d0c8d2-29dc-471f-8c60-61e0af4b9b22 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.678238] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1318998, 'name': ReconfigVM_Task, 'duration_secs': 0.234516} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.681483] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Reconfigured VM instance instance-0000005a to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1012.686733] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b21a0f4d-7f0e-410f-8995-06a6f51c4195 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.703535] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1318999, 'name': CreateVM_Task, 'duration_secs': 0.38099} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.704954] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1012.705320] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1012.705320] env[62521]: value = "task-1319000" [ 1012.705320] env[62521]: _type = "Task" [ 1012.705320] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.706037] env[62521]: DEBUG oslo_concurrency.lockutils [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.706177] env[62521]: DEBUG oslo_concurrency.lockutils [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.706460] env[62521]: DEBUG oslo_concurrency.lockutils [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1012.709625] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f29367e6-bf10-406a-b660-e5c4a22968ec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.713811] env[62521]: DEBUG oslo_vmware.api [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318994, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.716735] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1012.716735] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d3da4b-5296-7afc-6181-8228a1ca900a" [ 1012.716735] env[62521]: _type = "Task" [ 1012.716735] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.722941] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319000, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.728018] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d3da4b-5296-7afc-6181-8228a1ca900a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.101605] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e0cbbd-e418-4d93-9e14-f4bed0a7b3ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.109542] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd588ec-4366-41b4-9f91-2d96e0e46793 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.144156] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e73eb04-566e-449c-8e62-48392b949309 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.152250] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38e4874-648f-4078-b2e8-5c1cde31051d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.165783] env[62521]: DEBUG nova.compute.provider_tree [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.206126] env[62521]: DEBUG oslo_vmware.api [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1318994, 'name': PowerOffVM_Task, 'duration_secs': 2.134961} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.206674] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1013.206857] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1013.207170] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8dcd8cf1-c2eb-4506-bd69-33ab47935eec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.215927] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319000, 'name': ReconfigVM_Task, 'duration_secs': 0.182763} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.216225] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282168', 'volume_id': '510f1170-3625-496a-9477-442b0e639a96', 'name': 'volume-510f1170-3625-496a-9477-442b0e639a96', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7cdeed76-e5b3-4959-b6a6-610d23233e1b', 'attached_at': '', 'detached_at': '', 'volume_id': '510f1170-3625-496a-9477-442b0e639a96', 'serial': '510f1170-3625-496a-9477-442b0e639a96'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1013.216512] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1013.217254] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d77063c-88fb-4875-98d8-fcdf1080a1fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.227389] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d3da4b-5296-7afc-6181-8228a1ca900a, 'name': SearchDatastore_Task, 'duration_secs': 0.01313} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.229306] env[62521]: DEBUG oslo_concurrency.lockutils [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.229556] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.229769] env[62521]: DEBUG oslo_concurrency.lockutils [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.229918] env[62521]: DEBUG oslo_concurrency.lockutils [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.230110] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.230395] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1013.230614] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95af217f-5e6c-4c37-ba33-d0ef74ecfeaf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.232148] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d87b944a-386d-4ead-af78-85df168b539c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.240642] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.240822] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1013.241482] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f034a2c-8f10-44c8-86c6-5a5c51606c5e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.246578] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1013.246578] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529e85a3-4067-4c63-45a9-2cdaf609e44d" [ 1013.246578] env[62521]: _type = "Task" [ 1013.246578] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.257012] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529e85a3-4067-4c63-45a9-2cdaf609e44d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.282703] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1013.282953] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1013.283163] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleting the datastore file [datastore1] f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.283501] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9474cf89-972c-45fb-9dca-7126053c713c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.290432] env[62521]: DEBUG oslo_vmware.api [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1013.290432] env[62521]: value = "task-1319003" [ 1013.290432] env[62521]: _type = "Task" [ 1013.290432] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.296837] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1013.296996] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1013.297347] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleting the datastore file [datastore2] 7cdeed76-e5b3-4959-b6a6-610d23233e1b {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.300318] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-541af0cb-6e6b-4501-b222-51c79f0fa640 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.302068] env[62521]: DEBUG oslo_vmware.api [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319003, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.305729] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1013.305729] env[62521]: value = "task-1319004" [ 1013.305729] env[62521]: _type = "Task" [ 1013.305729] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.313118] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319004, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.314068] env[62521]: DEBUG nova.network.neutron [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance_info_cache with network_info: [{"id": "9e954d34-4424-459c-a022-fc039668c616", "address": "fa:16:3e:90:66:a0", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e954d34-44", "ovs_interfaceid": "9e954d34-4424-459c-a022-fc039668c616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.588316] env[62521]: DEBUG nova.compute.manager [req-34415cdb-1348-4c22-bfbe-c9832783c3f8 req-f3584c95-473b-446e-873d-317d534435dc service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Received event network-changed-75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1013.588922] env[62521]: DEBUG nova.compute.manager [req-34415cdb-1348-4c22-bfbe-c9832783c3f8 req-f3584c95-473b-446e-873d-317d534435dc service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Refreshing instance network info cache due to event network-changed-75ffddaa-cba0-4286-84c8-454e40d5bb4f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1013.589311] env[62521]: DEBUG oslo_concurrency.lockutils [req-34415cdb-1348-4c22-bfbe-c9832783c3f8 req-f3584c95-473b-446e-873d-317d534435dc service nova] Acquiring lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.589680] env[62521]: DEBUG oslo_concurrency.lockutils [req-34415cdb-1348-4c22-bfbe-c9832783c3f8 req-f3584c95-473b-446e-873d-317d534435dc service nova] Acquired lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.590044] env[62521]: DEBUG nova.network.neutron [req-34415cdb-1348-4c22-bfbe-c9832783c3f8 req-f3584c95-473b-446e-873d-317d534435dc service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Refreshing network info cache for port 75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1013.655022] env[62521]: DEBUG nova.network.neutron [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Successfully updated port: f5f1a879-9bd4-43e6-815a-307ae9cff9e7 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.668480] env[62521]: DEBUG nova.scheduler.client.report [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.756642] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529e85a3-4067-4c63-45a9-2cdaf609e44d, 'name': SearchDatastore_Task, 'duration_secs': 0.009599} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.757481] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34a7e787-fb9d-4b8d-8071-7e59f6e4e485 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.762396] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1013.762396] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f8f4af-dae9-0a71-0264-6086858c50a3" [ 1013.762396] env[62521]: _type = "Task" [ 1013.762396] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.769520] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f8f4af-dae9-0a71-0264-6086858c50a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.798069] env[62521]: DEBUG oslo_vmware.api [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319003, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152537} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.798332] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.798535] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1013.798729] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1013.798912] env[62521]: INFO nova.compute.manager [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Took 3.13 seconds to destroy the instance on the hypervisor. [ 1013.799182] env[62521]: DEBUG oslo.service.loopingcall [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.799378] env[62521]: DEBUG nova.compute.manager [-] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1013.799474] env[62521]: DEBUG nova.network.neutron [-] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1013.813446] env[62521]: DEBUG oslo_vmware.api [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319004, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149184} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.813833] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.813833] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1013.813974] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1013.814247] env[62521]: INFO nova.compute.manager [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Took 2.31 seconds to destroy the instance on the hypervisor. [ 1013.814485] env[62521]: DEBUG oslo.service.loopingcall [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.814670] env[62521]: DEBUG nova.compute.manager [-] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1013.814758] env[62521]: DEBUG nova.network.neutron [-] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1013.816364] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.156595] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.156595] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.156595] env[62521]: DEBUG nova.network.neutron [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.175534] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.289s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.176096] env[62521]: DEBUG nova.compute.manager [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1014.182024] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.668s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.182790] env[62521]: INFO nova.compute.claims [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.257075] env[62521]: DEBUG nova.compute.manager [req-72a7de8d-02db-4aba-a7ff-be782639ea86 req-42c60949-412c-482f-a797-0ee5e6a56ae4 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Received event network-vif-plugged-f5f1a879-9bd4-43e6-815a-307ae9cff9e7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.257345] env[62521]: DEBUG oslo_concurrency.lockutils [req-72a7de8d-02db-4aba-a7ff-be782639ea86 req-42c60949-412c-482f-a797-0ee5e6a56ae4 service nova] Acquiring lock "b8158051-98b6-47d8-84c5-d1d56515afe8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.257564] env[62521]: DEBUG oslo_concurrency.lockutils [req-72a7de8d-02db-4aba-a7ff-be782639ea86 req-42c60949-412c-482f-a797-0ee5e6a56ae4 service nova] Lock "b8158051-98b6-47d8-84c5-d1d56515afe8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.257830] env[62521]: DEBUG oslo_concurrency.lockutils [req-72a7de8d-02db-4aba-a7ff-be782639ea86 req-42c60949-412c-482f-a797-0ee5e6a56ae4 service nova] Lock "b8158051-98b6-47d8-84c5-d1d56515afe8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.257952] env[62521]: DEBUG nova.compute.manager [req-72a7de8d-02db-4aba-a7ff-be782639ea86 req-42c60949-412c-482f-a797-0ee5e6a56ae4 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] No waiting events found dispatching network-vif-plugged-f5f1a879-9bd4-43e6-815a-307ae9cff9e7 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.258068] env[62521]: WARNING nova.compute.manager [req-72a7de8d-02db-4aba-a7ff-be782639ea86 req-42c60949-412c-482f-a797-0ee5e6a56ae4 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Received unexpected event network-vif-plugged-f5f1a879-9bd4-43e6-815a-307ae9cff9e7 for instance with vm_state active and task_state None. [ 1014.272582] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f8f4af-dae9-0a71-0264-6086858c50a3, 'name': SearchDatastore_Task, 'duration_secs': 0.008911} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.272845] env[62521]: DEBUG oslo_concurrency.lockutils [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.273109] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] aa266be6-2bf1-445f-a968-2c144d30ecfa/aa266be6-2bf1-445f-a968-2c144d30ecfa.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1014.273369] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2fe3d2d8-e540-4b0c-9d24-37d5f5c1aa1e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.281490] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1014.281490] env[62521]: value = "task-1319005" [ 1014.281490] env[62521]: _type = "Task" [ 1014.281490] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.290130] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319005, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.313769] env[62521]: DEBUG nova.network.neutron [req-34415cdb-1348-4c22-bfbe-c9832783c3f8 req-f3584c95-473b-446e-873d-317d534435dc service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updated VIF entry in instance network info cache for port 75ffddaa-cba0-4286-84c8-454e40d5bb4f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1014.314185] env[62521]: DEBUG nova.network.neutron [req-34415cdb-1348-4c22-bfbe-c9832783c3f8 req-f3584c95-473b-446e-873d-317d534435dc service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updating instance_info_cache with network_info: [{"id": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "address": "fa:16:3e:82:b0:68", "network": {"id": "e84e7dd3-8f72-4865-86ba-63a68f5178cf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1633868122-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f513dc5025974607a90e67621154bccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75ffddaa-cb", "ovs_interfaceid": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.538639] env[62521]: DEBUG nova.network.neutron [-] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.634594] env[62521]: DEBUG nova.network.neutron [-] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.687350] env[62521]: DEBUG nova.compute.utils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.691172] env[62521]: DEBUG nova.compute.manager [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1014.691361] env[62521]: DEBUG nova.network.neutron [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1014.695595] env[62521]: WARNING nova.network.neutron [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] 67bdad13-8ff4-4bf8-b255-d0a285416fbe already exists in list: networks containing: ['67bdad13-8ff4-4bf8-b255-d0a285416fbe']. ignoring it [ 1014.761805] env[62521]: DEBUG nova.policy [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fa25ccc8dc064a3683b2dd1ae3844c6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a23c28c0b9fe4ab09d02186797bd73d1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1014.791391] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319005, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482033} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.792052] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] aa266be6-2bf1-445f-a968-2c144d30ecfa/aa266be6-2bf1-445f-a968-2c144d30ecfa.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1014.792052] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1014.792192] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-80edbb17-8118-4674-a97f-0f9ecd487ce9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.799030] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1014.799030] env[62521]: value = "task-1319006" [ 1014.799030] env[62521]: _type = "Task" [ 1014.799030] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.807331] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319006, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.816967] env[62521]: DEBUG oslo_concurrency.lockutils [req-34415cdb-1348-4c22-bfbe-c9832783c3f8 req-f3584c95-473b-446e-873d-317d534435dc service nova] Releasing lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.006956] env[62521]: DEBUG nova.network.neutron [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updating instance_info_cache with network_info: [{"id": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "address": "fa:16:3e:37:05:5a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b0cfbc-ec", "ovs_interfaceid": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f5f1a879-9bd4-43e6-815a-307ae9cff9e7", "address": "fa:16:3e:3a:89:5e", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5f1a879-9b", "ovs_interfaceid": "f5f1a879-9bd4-43e6-815a-307ae9cff9e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.044784] env[62521]: INFO nova.compute.manager [-] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Took 1.24 seconds to deallocate network for instance. [ 1015.136728] env[62521]: INFO nova.compute.manager [-] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Took 1.32 seconds to deallocate network for instance. [ 1015.162051] env[62521]: DEBUG nova.network.neutron [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Successfully created port: b7d04538-d173-4606-83a1-3c5b32ee6357 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1015.198216] env[62521]: DEBUG nova.compute.manager [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1015.309054] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319006, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065642} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.309585] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1015.310405] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb563a2-5981-4e14-8301-b1d9edc7c2e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.336468] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] aa266be6-2bf1-445f-a968-2c144d30ecfa/aa266be6-2bf1-445f-a968-2c144d30ecfa.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.340989] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e63df7e0-fc08-4c5e-ab5a-4fdffb16b7d8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.357811] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f000b258-c22c-49ee-b4ac-9caaf5d67588 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.384455] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance '11e61e72-a311-4c43-bf53-df82ad9d70f7' progress to 0 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1015.388496] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1015.388496] env[62521]: value = "task-1319007" [ 1015.388496] env[62521]: _type = "Task" [ 1015.388496] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.402177] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319007, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.419246] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c937e5-99e3-4095-ba5f-c6607dddf7e2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.426596] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03d2e52-9d78-43a8-b9bf-64a9ce24224f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.457428] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4055044f-0592-447a-998a-7c3dd2b11028 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.464707] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d7e515-0d71-4657-97f2-d4888cc874c8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.477848] env[62521]: DEBUG nova.compute.provider_tree [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1015.510192] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.510908] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.511146] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.512180] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19871869-bd16-4b3a-b12f-834e1c278fc4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.531372] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.531661] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.531852] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.532083] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.532255] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.532439] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.532709] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.532892] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.533094] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.533271] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.533453] env[62521]: DEBUG nova.virt.hardware [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.540037] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Reconfiguring VM to attach interface {{(pid=62521) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1015.540759] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da9ef958-b8bb-40df-8a1a-a89464238015 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.553250] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.559180] env[62521]: DEBUG oslo_vmware.api [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1015.559180] env[62521]: value = "task-1319008" [ 1015.559180] env[62521]: _type = "Task" [ 1015.559180] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.567545] env[62521]: DEBUG oslo_vmware.api [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319008, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.685676] env[62521]: INFO nova.compute.manager [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Took 0.55 seconds to detach 1 volumes for instance. [ 1015.892407] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.892407] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af9ce631-3ac6-4a87-bb08-78bd743c7edb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.902848] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319007, 'name': ReconfigVM_Task, 'duration_secs': 0.311718} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.904079] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Reconfigured VM instance instance-0000005d to attach disk [datastore1] aa266be6-2bf1-445f-a968-2c144d30ecfa/aa266be6-2bf1-445f-a968-2c144d30ecfa.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.904792] env[62521]: DEBUG oslo_vmware.api [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1015.904792] env[62521]: value = "task-1319009" [ 1015.904792] env[62521]: _type = "Task" [ 1015.904792] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.904996] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dcfe23bf-1028-4b06-8d2b-b8d6dc1f6b31 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.916196] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1015.916421] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance '11e61e72-a311-4c43-bf53-df82ad9d70f7' progress to 17 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1015.920047] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1015.920047] env[62521]: value = "task-1319010" [ 1015.920047] env[62521]: _type = "Task" [ 1015.920047] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.928695] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319010, 'name': Rename_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.003948] env[62521]: ERROR nova.scheduler.client.report [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [req-10b24cc5-c10f-4c83-9db5-d8e146afedf7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-10b24cc5-c10f-4c83-9db5-d8e146afedf7"}]} [ 1016.023074] env[62521]: DEBUG nova.scheduler.client.report [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1016.040303] env[62521]: DEBUG nova.scheduler.client.report [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1016.040303] env[62521]: DEBUG nova.compute.provider_tree [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1016.053818] env[62521]: DEBUG nova.scheduler.client.report [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1016.071495] env[62521]: DEBUG oslo_vmware.api [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319008, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.078449] env[62521]: DEBUG nova.scheduler.client.report [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1016.192317] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.207087] env[62521]: DEBUG nova.compute.manager [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1016.236884] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.237168] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.237361] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.237565] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.237731] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.237910] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.238137] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.238322] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.238497] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.238665] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.238840] env[62521]: DEBUG nova.virt.hardware [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.239761] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0097f2b8-66cd-454a-afcf-b7f274c27a2e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.248935] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3598b59-ae00-48a5-bd8f-9b16eb73df7a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.277547] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ced8185-7aa3-4fde-9236-683e8cc42652 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.285997] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21e8c33-7975-49ca-921f-1c355e566094 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.321052] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a77f5e-782e-4596-9d96-b43bacc0a84e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.328471] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3f661a-f1d0-4e20-8fb7-94e41369cee5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.342326] env[62521]: DEBUG nova.compute.provider_tree [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1016.423127] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.423419] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.423586] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.423772] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.423925] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.424088] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.424295] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.424609] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.424751] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.424834] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.425081] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.430177] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a37eb19-8aab-45cd-a0ce-94f5e1ae9194 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.449299] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319010, 'name': Rename_Task, 'duration_secs': 0.182066} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.450538] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1016.450829] env[62521]: DEBUG oslo_vmware.api [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1016.450829] env[62521]: value = "task-1319011" [ 1016.450829] env[62521]: _type = "Task" [ 1016.450829] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.451057] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fab1cd76-5d79-4b55-b449-090f0dc25dfc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.460101] env[62521]: DEBUG oslo_vmware.api [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319011, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.461208] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1016.461208] env[62521]: value = "task-1319012" [ 1016.461208] env[62521]: _type = "Task" [ 1016.461208] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.468990] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319012, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.500157] env[62521]: DEBUG nova.compute.manager [req-1658c140-5594-4eef-9331-632ab919af31 req-81b8d602-0e22-4f20-81b6-c7fc0f250ac7 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Received event network-changed-75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.500393] env[62521]: DEBUG nova.compute.manager [req-1658c140-5594-4eef-9331-632ab919af31 req-81b8d602-0e22-4f20-81b6-c7fc0f250ac7 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Refreshing instance network info cache due to event network-changed-75ffddaa-cba0-4286-84c8-454e40d5bb4f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1016.500633] env[62521]: DEBUG oslo_concurrency.lockutils [req-1658c140-5594-4eef-9331-632ab919af31 req-81b8d602-0e22-4f20-81b6-c7fc0f250ac7 service nova] Acquiring lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.500787] env[62521]: DEBUG oslo_concurrency.lockutils [req-1658c140-5594-4eef-9331-632ab919af31 req-81b8d602-0e22-4f20-81b6-c7fc0f250ac7 service nova] Acquired lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.500974] env[62521]: DEBUG nova.network.neutron [req-1658c140-5594-4eef-9331-632ab919af31 req-81b8d602-0e22-4f20-81b6-c7fc0f250ac7 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Refreshing network info cache for port 75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1016.570279] env[62521]: DEBUG oslo_vmware.api [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319008, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.881006] env[62521]: DEBUG nova.scheduler.client.report [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 126 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1016.881304] env[62521]: DEBUG nova.compute.provider_tree [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 126 to 127 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1016.881492] env[62521]: DEBUG nova.compute.provider_tree [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1016.963586] env[62521]: DEBUG oslo_vmware.api [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319011, 'name': ReconfigVM_Task, 'duration_secs': 0.198184} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.965892] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance '11e61e72-a311-4c43-bf53-df82ad9d70f7' progress to 33 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1016.976606] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319012, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.070551] env[62521]: DEBUG oslo_vmware.api [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319008, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.200961] env[62521]: DEBUG nova.network.neutron [req-1658c140-5594-4eef-9331-632ab919af31 req-81b8d602-0e22-4f20-81b6-c7fc0f250ac7 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updated VIF entry in instance network info cache for port 75ffddaa-cba0-4286-84c8-454e40d5bb4f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1017.201389] env[62521]: DEBUG nova.network.neutron [req-1658c140-5594-4eef-9331-632ab919af31 req-81b8d602-0e22-4f20-81b6-c7fc0f250ac7 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updating instance_info_cache with network_info: [{"id": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "address": "fa:16:3e:82:b0:68", "network": {"id": "e84e7dd3-8f72-4865-86ba-63a68f5178cf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1633868122-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f513dc5025974607a90e67621154bccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75ffddaa-cb", "ovs_interfaceid": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.387268] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.206s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.387857] env[62521]: DEBUG nova.compute.manager [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1017.390690] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.837s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.393671] env[62521]: DEBUG nova.objects.instance [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lazy-loading 'resources' on Instance uuid f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.415418] env[62521]: DEBUG nova.network.neutron [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Successfully updated port: b7d04538-d173-4606-83a1-3c5b32ee6357 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1017.472023] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1017.472322] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1017.472525] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1017.472741] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1017.472901] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1017.473069] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1017.473293] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1017.473452] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1017.473620] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1017.473780] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1017.473949] env[62521]: DEBUG nova.virt.hardware [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.480142] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Reconfiguring VM instance instance-0000004c to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1017.484179] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-734292f4-ca6f-4ecc-9c45-98877f15875b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.497545] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319012, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.500830] env[62521]: DEBUG nova.compute.manager [req-77605129-9141-4184-b6dc-8a76ec1eac09 req-f586644f-b19b-451a-9361-c5cff36a8bac service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Received event network-changed-f5f1a879-9bd4-43e6-815a-307ae9cff9e7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.501070] env[62521]: DEBUG nova.compute.manager [req-77605129-9141-4184-b6dc-8a76ec1eac09 req-f586644f-b19b-451a-9361-c5cff36a8bac service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Refreshing instance network info cache due to event network-changed-f5f1a879-9bd4-43e6-815a-307ae9cff9e7. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1017.501301] env[62521]: DEBUG oslo_concurrency.lockutils [req-77605129-9141-4184-b6dc-8a76ec1eac09 req-f586644f-b19b-451a-9361-c5cff36a8bac service nova] Acquiring lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.501493] env[62521]: DEBUG oslo_concurrency.lockutils [req-77605129-9141-4184-b6dc-8a76ec1eac09 req-f586644f-b19b-451a-9361-c5cff36a8bac service nova] Acquired lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.501703] env[62521]: DEBUG nova.network.neutron [req-77605129-9141-4184-b6dc-8a76ec1eac09 req-f586644f-b19b-451a-9361-c5cff36a8bac service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Refreshing network info cache for port f5f1a879-9bd4-43e6-815a-307ae9cff9e7 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1017.508972] env[62521]: DEBUG oslo_vmware.api [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1017.508972] env[62521]: value = "task-1319013" [ 1017.508972] env[62521]: _type = "Task" [ 1017.508972] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.509554] env[62521]: DEBUG nova.compute.manager [req-4246349b-0269-4e4b-9df7-1f7cf8a894ac req-d25feb9c-4689-4e6e-82fa-6625767f967c service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Received event network-vif-plugged-b7d04538-d173-4606-83a1-3c5b32ee6357 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.509777] env[62521]: DEBUG oslo_concurrency.lockutils [req-4246349b-0269-4e4b-9df7-1f7cf8a894ac req-d25feb9c-4689-4e6e-82fa-6625767f967c service nova] Acquiring lock "83fe58c9-920b-422d-be08-e5d53bf551d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.509988] env[62521]: DEBUG oslo_concurrency.lockutils [req-4246349b-0269-4e4b-9df7-1f7cf8a894ac req-d25feb9c-4689-4e6e-82fa-6625767f967c service nova] Lock "83fe58c9-920b-422d-be08-e5d53bf551d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.510174] env[62521]: DEBUG oslo_concurrency.lockutils [req-4246349b-0269-4e4b-9df7-1f7cf8a894ac req-d25feb9c-4689-4e6e-82fa-6625767f967c service nova] Lock "83fe58c9-920b-422d-be08-e5d53bf551d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.510348] env[62521]: DEBUG nova.compute.manager [req-4246349b-0269-4e4b-9df7-1f7cf8a894ac req-d25feb9c-4689-4e6e-82fa-6625767f967c service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] No waiting events found dispatching network-vif-plugged-b7d04538-d173-4606-83a1-3c5b32ee6357 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1017.510639] env[62521]: WARNING nova.compute.manager [req-4246349b-0269-4e4b-9df7-1f7cf8a894ac req-d25feb9c-4689-4e6e-82fa-6625767f967c service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Received unexpected event network-vif-plugged-b7d04538-d173-4606-83a1-3c5b32ee6357 for instance with vm_state building and task_state spawning. [ 1017.520548] env[62521]: DEBUG oslo_vmware.api [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319013, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.571555] env[62521]: DEBUG oslo_vmware.api [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319008, 'name': ReconfigVM_Task, 'duration_secs': 1.673732} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.572114] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.572363] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Reconfigured VM to attach interface {{(pid=62521) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1017.704584] env[62521]: DEBUG oslo_concurrency.lockutils [req-1658c140-5594-4eef-9331-632ab919af31 req-81b8d602-0e22-4f20-81b6-c7fc0f250ac7 service nova] Releasing lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.704934] env[62521]: DEBUG nova.compute.manager [req-1658c140-5594-4eef-9331-632ab919af31 req-81b8d602-0e22-4f20-81b6-c7fc0f250ac7 service nova] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Received event network-vif-deleted-56ec9ce6-ead9-4a28-b282-1bfeba30bdf6 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.894443] env[62521]: DEBUG nova.compute.utils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1017.899958] env[62521]: DEBUG nova.compute.manager [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1017.900304] env[62521]: DEBUG nova.network.neutron [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1017.919480] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquiring lock "refresh_cache-83fe58c9-920b-422d-be08-e5d53bf551d3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.919640] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquired lock "refresh_cache-83fe58c9-920b-422d-be08-e5d53bf551d3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.919814] env[62521]: DEBUG nova.network.neutron [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1017.974619] env[62521]: DEBUG oslo_vmware.api [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319012, 'name': PowerOnVM_Task, 'duration_secs': 1.311421} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.976407] env[62521]: DEBUG nova.policy [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e602eba085748d59dc453e05c98e39e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ab001785c25414abbaa4bfe3a6a0bb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1017.978094] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1017.978520] env[62521]: DEBUG nova.compute.manager [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.979457] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8c157b-09fa-4546-9ea6-c87b4461a6c6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.021055] env[62521]: DEBUG oslo_vmware.api [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319013, 'name': ReconfigVM_Task, 'duration_secs': 0.152401} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.021374] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Reconfigured VM instance instance-0000004c to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1018.022182] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c262fb-ecdc-4a05-9907-8b962a64e852 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.052034] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 11e61e72-a311-4c43-bf53-df82ad9d70f7/11e61e72-a311-4c43-bf53-df82ad9d70f7.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.058014] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a32389f3-a079-4aff-bf2d-8dfa947c2b00 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.078283] env[62521]: DEBUG oslo_concurrency.lockutils [None req-edfb11f8-751d-4880-9ad0-7be2ad653996 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-b8158051-98b6-47d8-84c5-d1d56515afe8-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.385s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.081299] env[62521]: DEBUG oslo_vmware.api [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1018.081299] env[62521]: value = "task-1319014" [ 1018.081299] env[62521]: _type = "Task" [ 1018.081299] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.090769] env[62521]: DEBUG oslo_vmware.api [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319014, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.137076] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde34c4f-3423-494e-b215-6d8a357b876f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.150568] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22c1c81-d839-4044-8d57-6ed0b1fc4cac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.200762] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63e81d5-cdaa-4294-a8f6-34403eea391c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.211276] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27deb142-80ca-42b7-bb77-79e03701442b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.227085] env[62521]: DEBUG nova.compute.provider_tree [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.334681] env[62521]: DEBUG nova.network.neutron [req-77605129-9141-4184-b6dc-8a76ec1eac09 req-f586644f-b19b-451a-9361-c5cff36a8bac service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updated VIF entry in instance network info cache for port f5f1a879-9bd4-43e6-815a-307ae9cff9e7. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1018.335163] env[62521]: DEBUG nova.network.neutron [req-77605129-9141-4184-b6dc-8a76ec1eac09 req-f586644f-b19b-451a-9361-c5cff36a8bac service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updating instance_info_cache with network_info: [{"id": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "address": "fa:16:3e:37:05:5a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b0cfbc-ec", "ovs_interfaceid": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f5f1a879-9bd4-43e6-815a-307ae9cff9e7", "address": "fa:16:3e:3a:89:5e", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5f1a879-9b", "ovs_interfaceid": "f5f1a879-9bd4-43e6-815a-307ae9cff9e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.400672] env[62521]: DEBUG nova.compute.manager [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1018.455880] env[62521]: DEBUG nova.network.neutron [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1018.459171] env[62521]: DEBUG nova.network.neutron [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Successfully created port: a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1018.503961] env[62521]: DEBUG oslo_concurrency.lockutils [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.588293] env[62521]: DEBUG nova.network.neutron [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Updating instance_info_cache with network_info: [{"id": "b7d04538-d173-4606-83a1-3c5b32ee6357", "address": "fa:16:3e:09:44:4b", "network": {"id": "51fbdbaf-b547-46b6-bf27-e4333163b9cc", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-218491247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c28c0b9fe4ab09d02186797bd73d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7d04538-d1", "ovs_interfaceid": "b7d04538-d173-4606-83a1-3c5b32ee6357", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.593142] env[62521]: DEBUG oslo_vmware.api [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319014, 'name': ReconfigVM_Task, 'duration_secs': 0.26351} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.593673] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 11e61e72-a311-4c43-bf53-df82ad9d70f7/11e61e72-a311-4c43-bf53-df82ad9d70f7.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.593899] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance '11e61e72-a311-4c43-bf53-df82ad9d70f7' progress to 50 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1018.730413] env[62521]: DEBUG nova.scheduler.client.report [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.838214] env[62521]: DEBUG oslo_concurrency.lockutils [req-77605129-9141-4184-b6dc-8a76ec1eac09 req-f586644f-b19b-451a-9361-c5cff36a8bac service nova] Releasing lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.838519] env[62521]: DEBUG nova.compute.manager [req-77605129-9141-4184-b6dc-8a76ec1eac09 req-f586644f-b19b-451a-9361-c5cff36a8bac service nova] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Received event network-vif-deleted-45c9722b-0936-4e02-8f69-17a30523490b {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.916255] env[62521]: DEBUG nova.compute.manager [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Received event network-changed-75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.916636] env[62521]: DEBUG nova.compute.manager [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Refreshing instance network info cache due to event network-changed-75ffddaa-cba0-4286-84c8-454e40d5bb4f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1018.916932] env[62521]: DEBUG oslo_concurrency.lockutils [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] Acquiring lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.917044] env[62521]: DEBUG oslo_concurrency.lockutils [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] Acquired lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.917217] env[62521]: DEBUG nova.network.neutron [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Refreshing network info cache for port 75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1019.095026] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Releasing lock "refresh_cache-83fe58c9-920b-422d-be08-e5d53bf551d3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.095352] env[62521]: DEBUG nova.compute.manager [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Instance network_info: |[{"id": "b7d04538-d173-4606-83a1-3c5b32ee6357", "address": "fa:16:3e:09:44:4b", "network": {"id": "51fbdbaf-b547-46b6-bf27-e4333163b9cc", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-218491247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c28c0b9fe4ab09d02186797bd73d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7d04538-d1", "ovs_interfaceid": "b7d04538-d173-4606-83a1-3c5b32ee6357", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1019.095862] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:44:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '62f28d75-4e6a-4ae5-b8b3-d0652ea26d08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7d04538-d173-4606-83a1-3c5b32ee6357', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1019.103740] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Creating folder: Project (a23c28c0b9fe4ab09d02186797bd73d1). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1019.105465] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1b6f3861-0c91-4d9c-944d-00821294543c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.107997] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0308f05-55bd-4eac-a0f4-4abc957052a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.130938] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015b355c-8bda-4205-913c-8b3cc5ea1d98 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.133729] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Created folder: Project (a23c28c0b9fe4ab09d02186797bd73d1) in parent group-v282025. [ 1019.133960] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Creating folder: Instances. Parent ref: group-v282170. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1019.134213] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4019b597-8f6c-411a-a002-1fc4521e3aee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.152014] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance '11e61e72-a311-4c43-bf53-df82ad9d70f7' progress to 67 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1019.159850] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Created folder: Instances in parent group-v282170. [ 1019.159850] env[62521]: DEBUG oslo.service.loopingcall [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.159850] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1019.159850] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-61aa8818-f4bc-402c-be61-069f720d0301 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.177977] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.177977] env[62521]: value = "task-1319017" [ 1019.177977] env[62521]: _type = "Task" [ 1019.177977] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.188078] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319017, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.235982] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.845s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.238364] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.046s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.238604] env[62521]: DEBUG nova.objects.instance [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lazy-loading 'resources' on Instance uuid 7cdeed76-e5b3-4959-b6a6-610d23233e1b {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.262163] env[62521]: INFO nova.scheduler.client.report [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted allocations for instance f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d [ 1019.410460] env[62521]: DEBUG nova.compute.manager [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1019.439095] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1019.439386] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1019.439553] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1019.439742] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1019.439892] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1019.440058] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1019.440277] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1019.440441] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1019.440614] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1019.440782] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1019.440957] env[62521]: DEBUG nova.virt.hardware [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1019.441871] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e8c916-49d1-4a7a-8349-fe55c25ac07c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.450278] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03237a9d-be76-4975-9a77-4c318af63429 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.546851] env[62521]: DEBUG oslo_concurrency.lockutils [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquiring lock "31f926a8-777d-4882-afd6-1e4dba3f4e11" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.546980] env[62521]: DEBUG oslo_concurrency.lockutils [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lock "31f926a8-777d-4882-afd6-1e4dba3f4e11" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.547211] env[62521]: DEBUG oslo_concurrency.lockutils [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquiring lock "31f926a8-777d-4882-afd6-1e4dba3f4e11-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.547434] env[62521]: DEBUG oslo_concurrency.lockutils [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lock "31f926a8-777d-4882-afd6-1e4dba3f4e11-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.547640] env[62521]: DEBUG oslo_concurrency.lockutils [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lock "31f926a8-777d-4882-afd6-1e4dba3f4e11-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.551840] env[62521]: INFO nova.compute.manager [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Terminating instance [ 1019.553740] env[62521]: DEBUG nova.compute.manager [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1019.553949] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1019.554793] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbab7cb0-5951-4e47-92e7-238734cb03d3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.563611] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1019.564736] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ce2650a-36e6-4651-91c9-6a9c177392de {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.569510] env[62521]: DEBUG oslo_vmware.api [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1019.569510] env[62521]: value = "task-1319018" [ 1019.569510] env[62521]: _type = "Task" [ 1019.569510] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.577617] env[62521]: DEBUG oslo_vmware.api [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1319018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.626683] env[62521]: DEBUG nova.network.neutron [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updated VIF entry in instance network info cache for port 75ffddaa-cba0-4286-84c8-454e40d5bb4f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1019.627093] env[62521]: DEBUG nova.network.neutron [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updating instance_info_cache with network_info: [{"id": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "address": "fa:16:3e:82:b0:68", "network": {"id": "e84e7dd3-8f72-4865-86ba-63a68f5178cf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1633868122-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f513dc5025974607a90e67621154bccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75ffddaa-cb", "ovs_interfaceid": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.691688] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319017, 'name': CreateVM_Task, 'duration_secs': 0.316628} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.691688] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1019.692297] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.692469] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.692927] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1019.693620] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66e82a58-9117-4f8f-aaf0-b5d5d22e8956 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.696203] env[62521]: DEBUG nova.network.neutron [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Port 9e954d34-4424-459c-a022-fc039668c616 binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1019.701508] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for the task: (returnval){ [ 1019.701508] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b8d92c-5100-2074-bedc-6dec2bfc5d8d" [ 1019.701508] env[62521]: _type = "Task" [ 1019.701508] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.710483] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b8d92c-5100-2074-bedc-6dec2bfc5d8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.768945] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0493cae7-c52d-4610-8ea8-97341d7a35d1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.109s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.806444] env[62521]: DEBUG oslo_concurrency.lockutils [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "interface-b8158051-98b6-47d8-84c5-d1d56515afe8-f5f1a879-9bd4-43e6-815a-307ae9cff9e7" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.806764] env[62521]: DEBUG oslo_concurrency.lockutils [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-b8158051-98b6-47d8-84c5-d1d56515afe8-f5f1a879-9bd4-43e6-815a-307ae9cff9e7" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.924356] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c12a7226-0322-4dae-8acb-f2077c86b74d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.933985] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946244be-46ff-44c0-9b2b-96f6e898900a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.967053] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4629c524-5f7b-43ec-ba1a-44418d76f1c8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.974614] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73029b8-e3ca-4faf-9e70-697712b02bd5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.988392] env[62521]: DEBUG nova.compute.provider_tree [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.080207] env[62521]: DEBUG oslo_vmware.api [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1319018, 'name': PowerOffVM_Task, 'duration_secs': 0.217161} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.080500] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1020.080678] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1020.080937] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ddd90bfb-54ad-48e0-806a-949ffbba1c2c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.129911] env[62521]: DEBUG oslo_concurrency.lockutils [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] Releasing lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.130233] env[62521]: DEBUG nova.compute.manager [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Received event network-changed-75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.130421] env[62521]: DEBUG nova.compute.manager [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Refreshing instance network info cache due to event network-changed-75ffddaa-cba0-4286-84c8-454e40d5bb4f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.130643] env[62521]: DEBUG oslo_concurrency.lockutils [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] Acquiring lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.130820] env[62521]: DEBUG oslo_concurrency.lockutils [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] Acquired lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.130957] env[62521]: DEBUG nova.network.neutron [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Refreshing network info cache for port 75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1020.146224] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1020.146476] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1020.146666] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Deleting the datastore file [datastore2] 31f926a8-777d-4882-afd6-1e4dba3f4e11 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.146932] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3622c337-e176-4467-a8cf-fa5c72ebf94b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.153196] env[62521]: DEBUG oslo_vmware.api [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for the task: (returnval){ [ 1020.153196] env[62521]: value = "task-1319020" [ 1020.153196] env[62521]: _type = "Task" [ 1020.153196] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.161285] env[62521]: DEBUG oslo_vmware.api [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1319020, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.190498] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "aa266be6-2bf1-445f-a968-2c144d30ecfa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.190498] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "aa266be6-2bf1-445f-a968-2c144d30ecfa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.190498] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "aa266be6-2bf1-445f-a968-2c144d30ecfa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.190498] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "aa266be6-2bf1-445f-a968-2c144d30ecfa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.190498] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "aa266be6-2bf1-445f-a968-2c144d30ecfa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.191576] env[62521]: INFO nova.compute.manager [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Terminating instance [ 1020.193759] env[62521]: DEBUG nova.compute.manager [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1020.194376] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.195506] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945c6789-67b4-434f-9c42-1a11a1f636fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.210345] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.211173] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46467479-220f-4f70-bafa-ba0279d7d2fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.216587] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b8d92c-5100-2074-bedc-6dec2bfc5d8d, 'name': SearchDatastore_Task, 'duration_secs': 0.009358} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.217416] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.217833] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1020.218236] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.218576] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.218912] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.219293] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e472177e-b93b-4a52-9726-dff54f56eedc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.223034] env[62521]: DEBUG oslo_vmware.api [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1020.223034] env[62521]: value = "task-1319021" [ 1020.223034] env[62521]: _type = "Task" [ 1020.223034] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.228724] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.231027] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1020.234402] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e682fea-40a6-4bbb-9dd9-e789d9049bbe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.237192] env[62521]: DEBUG oslo_vmware.api [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.241021] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for the task: (returnval){ [ 1020.241021] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fa0164-5c00-d176-bde9-89ddde7cde03" [ 1020.241021] env[62521]: _type = "Task" [ 1020.241021] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.251295] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fa0164-5c00-d176-bde9-89ddde7cde03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.314093] env[62521]: DEBUG oslo_concurrency.lockutils [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.314093] env[62521]: DEBUG oslo_concurrency.lockutils [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.314093] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50df8fff-c67b-4df2-be7c-e630564ee173 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.332548] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c671743-6bac-4d5b-aa9d-f50f9a24618c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.365868] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Reconfiguring VM to detach interface {{(pid=62521) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1020.366472] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3477dfa-b58c-4683-9de4-35d256f4183a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.391467] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1020.391467] env[62521]: value = "task-1319022" [ 1020.391467] env[62521]: _type = "Task" [ 1020.391467] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.392920] env[62521]: DEBUG nova.network.neutron [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Successfully updated port: a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.408646] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.492379] env[62521]: DEBUG nova.scheduler.client.report [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.663979] env[62521]: DEBUG oslo_vmware.api [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Task: {'id': task-1319020, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.330661} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.664267] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.664454] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1020.664636] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1020.664814] env[62521]: INFO nova.compute.manager [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1020.665111] env[62521]: DEBUG oslo.service.loopingcall [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.665496] env[62521]: DEBUG nova.compute.manager [-] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.665496] env[62521]: DEBUG nova.network.neutron [-] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1020.694260] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "60c23400-7eb2-4049-8a3b-77599098e334" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.694543] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "60c23400-7eb2-4049-8a3b-77599098e334" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.728818] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "11e61e72-a311-4c43-bf53-df82ad9d70f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.729071] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.729262] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.740569] env[62521]: DEBUG oslo_vmware.api [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319021, 'name': PowerOffVM_Task, 'duration_secs': 0.203705} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.741259] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1020.741442] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1020.741689] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d491b317-2e28-4b7a-ab11-c6a32e8b18c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.755109] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fa0164-5c00-d176-bde9-89ddde7cde03, 'name': SearchDatastore_Task, 'duration_secs': 0.009665} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.755872] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-946a050a-4379-414a-be25-be5e1cec3240 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.760650] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for the task: (returnval){ [ 1020.760650] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b5b23e-9958-5618-13ab-bdec8e6837fd" [ 1020.760650] env[62521]: _type = "Task" [ 1020.760650] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.768474] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b5b23e-9958-5618-13ab-bdec8e6837fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.833129] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1020.833384] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1020.833578] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleting the datastore file [datastore1] aa266be6-2bf1-445f-a968-2c144d30ecfa {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.833846] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19aa0f96-6367-4e35-be27-33a2081471f1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.842795] env[62521]: DEBUG oslo_vmware.api [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1020.842795] env[62521]: value = "task-1319024" [ 1020.842795] env[62521]: _type = "Task" [ 1020.842795] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.850524] env[62521]: DEBUG nova.network.neutron [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updated VIF entry in instance network info cache for port 75ffddaa-cba0-4286-84c8-454e40d5bb4f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1020.850918] env[62521]: DEBUG nova.network.neutron [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updating instance_info_cache with network_info: [{"id": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "address": "fa:16:3e:82:b0:68", "network": {"id": "e84e7dd3-8f72-4865-86ba-63a68f5178cf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1633868122-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f513dc5025974607a90e67621154bccd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75ffddaa-cb", "ovs_interfaceid": "75ffddaa-cba0-4286-84c8-454e40d5bb4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.852157] env[62521]: DEBUG oslo_vmware.api [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319024, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.902089] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.902089] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.902089] env[62521]: DEBUG nova.network.neutron [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1020.907052] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.931535] env[62521]: DEBUG nova.compute.manager [req-d2a59068-e303-4dce-be6f-f12e09f7cf9c req-378b219e-d5ef-48b1-94f1-99c99e94a4bc service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Received event network-vif-deleted-75ffddaa-cba0-4286-84c8-454e40d5bb4f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.931535] env[62521]: INFO nova.compute.manager [req-d2a59068-e303-4dce-be6f-f12e09f7cf9c req-378b219e-d5ef-48b1-94f1-99c99e94a4bc service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Neutron deleted interface 75ffddaa-cba0-4286-84c8-454e40d5bb4f; detaching it from the instance and deleting it from the info cache [ 1020.932100] env[62521]: DEBUG nova.network.neutron [req-d2a59068-e303-4dce-be6f-f12e09f7cf9c req-378b219e-d5ef-48b1-94f1-99c99e94a4bc service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.951590] env[62521]: DEBUG nova.compute.manager [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received event network-vif-plugged-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.951590] env[62521]: DEBUG oslo_concurrency.lockutils [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] Acquiring lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.951590] env[62521]: DEBUG oslo_concurrency.lockutils [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.951590] env[62521]: DEBUG oslo_concurrency.lockutils [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.951590] env[62521]: DEBUG nova.compute.manager [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] No waiting events found dispatching network-vif-plugged-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.951590] env[62521]: WARNING nova.compute.manager [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received unexpected event network-vif-plugged-a2f01aec-5d56-42ef-ac6c-6364f1571174 for instance with vm_state building and task_state spawning. [ 1020.951590] env[62521]: DEBUG nova.compute.manager [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received event network-changed-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.951590] env[62521]: DEBUG nova.compute.manager [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Refreshing instance network info cache due to event network-changed-a2f01aec-5d56-42ef-ac6c-6364f1571174. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.951877] env[62521]: DEBUG oslo_concurrency.lockutils [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] Acquiring lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.997084] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.759s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.999641] env[62521]: DEBUG oslo_concurrency.lockutils [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.496s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.000101] env[62521]: DEBUG nova.objects.instance [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62521) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1021.020468] env[62521]: INFO nova.scheduler.client.report [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted allocations for instance 7cdeed76-e5b3-4959-b6a6-610d23233e1b [ 1021.196654] env[62521]: DEBUG nova.compute.manager [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1021.271133] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b5b23e-9958-5618-13ab-bdec8e6837fd, 'name': SearchDatastore_Task, 'duration_secs': 0.009883} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.271421] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.271692] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 83fe58c9-920b-422d-be08-e5d53bf551d3/83fe58c9-920b-422d-be08-e5d53bf551d3.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1021.271951] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb86a3b4-9696-4619-bf4e-63bafd467a2c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.278265] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for the task: (returnval){ [ 1021.278265] env[62521]: value = "task-1319025" [ 1021.278265] env[62521]: _type = "Task" [ 1021.278265] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.285705] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319025, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.351787] env[62521]: DEBUG oslo_vmware.api [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319024, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143789} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.352073] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.352270] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1021.352471] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1021.352649] env[62521]: INFO nova.compute.manager [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1021.352891] env[62521]: DEBUG oslo.service.loopingcall [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1021.353351] env[62521]: DEBUG oslo_concurrency.lockutils [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] Releasing lock "refresh_cache-31f926a8-777d-4882-afd6-1e4dba3f4e11" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.353578] env[62521]: DEBUG nova.compute.manager [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Received event network-changed-b7d04538-d173-4606-83a1-3c5b32ee6357 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.353791] env[62521]: DEBUG nova.compute.manager [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Refreshing instance network info cache due to event network-changed-b7d04538-d173-4606-83a1-3c5b32ee6357. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1021.353950] env[62521]: DEBUG oslo_concurrency.lockutils [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] Acquiring lock "refresh_cache-83fe58c9-920b-422d-be08-e5d53bf551d3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.354108] env[62521]: DEBUG oslo_concurrency.lockutils [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] Acquired lock "refresh_cache-83fe58c9-920b-422d-be08-e5d53bf551d3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.354275] env[62521]: DEBUG nova.network.neutron [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Refreshing network info cache for port b7d04538-d173-4606-83a1-3c5b32ee6357 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1021.355462] env[62521]: DEBUG nova.compute.manager [-] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1021.355583] env[62521]: DEBUG nova.network.neutron [-] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1021.406955] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.410776] env[62521]: DEBUG nova.network.neutron [-] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.434604] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f71a6a0-b4b3-451c-99de-b311e93374a2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.443513] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba9f773-ae75-4aff-bd8f-191cac1007e3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.479364] env[62521]: DEBUG nova.compute.manager [req-d2a59068-e303-4dce-be6f-f12e09f7cf9c req-378b219e-d5ef-48b1-94f1-99c99e94a4bc service nova] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Detach interface failed, port_id=75ffddaa-cba0-4286-84c8-454e40d5bb4f, reason: Instance 31f926a8-777d-4882-afd6-1e4dba3f4e11 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1021.480684] env[62521]: DEBUG nova.network.neutron [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1021.528133] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e54b6f6-5857-433a-b27c-22cda5ba849f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "7cdeed76-e5b3-4959-b6a6-610d23233e1b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.023s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.720938] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.764110] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.764326] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.764519] env[62521]: DEBUG nova.network.neutron [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1021.789362] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319025, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.807181] env[62521]: DEBUG nova.network.neutron [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating instance_info_cache with network_info: [{"id": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "address": "fa:16:3e:dd:37:93", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2f01aec-5d", "ovs_interfaceid": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.905225] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.916873] env[62521]: INFO nova.compute.manager [-] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Took 1.25 seconds to deallocate network for instance. [ 1022.008157] env[62521]: DEBUG oslo_concurrency.lockutils [None req-723d904f-c759-4191-acb7-b02cba81c474 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.010111] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.290s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.011845] env[62521]: INFO nova.compute.claims [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1022.116165] env[62521]: DEBUG nova.network.neutron [-] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.137455] env[62521]: DEBUG nova.network.neutron [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Updated VIF entry in instance network info cache for port b7d04538-d173-4606-83a1-3c5b32ee6357. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1022.137794] env[62521]: DEBUG nova.network.neutron [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Updating instance_info_cache with network_info: [{"id": "b7d04538-d173-4606-83a1-3c5b32ee6357", "address": "fa:16:3e:09:44:4b", "network": {"id": "51fbdbaf-b547-46b6-bf27-e4333163b9cc", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-218491247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c28c0b9fe4ab09d02186797bd73d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7d04538-d1", "ovs_interfaceid": "b7d04538-d173-4606-83a1-3c5b32ee6357", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.290095] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319025, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.536598} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.290095] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 83fe58c9-920b-422d-be08-e5d53bf551d3/83fe58c9-920b-422d-be08-e5d53bf551d3.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1022.290095] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1022.290415] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-233d595a-f8c8-4afb-a9fc-0a779c16fdd6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.297029] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for the task: (returnval){ [ 1022.297029] env[62521]: value = "task-1319026" [ 1022.297029] env[62521]: _type = "Task" [ 1022.297029] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.304166] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319026, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.310185] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.310185] env[62521]: DEBUG nova.compute.manager [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Instance network_info: |[{"id": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "address": "fa:16:3e:dd:37:93", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2f01aec-5d", "ovs_interfaceid": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1022.312455] env[62521]: DEBUG oslo_concurrency.lockutils [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] Acquired lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.312645] env[62521]: DEBUG nova.network.neutron [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Refreshing network info cache for port a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1022.314570] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:37:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ec3f9e71-839a-429d-b211-d3dfc98ca4f6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a2f01aec-5d56-42ef-ac6c-6364f1571174', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.322803] env[62521]: DEBUG oslo.service.loopingcall [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.323588] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1022.324122] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-20efdaf4-8ff8-4905-995a-d32e735e232a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.345065] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.345065] env[62521]: value = "task-1319027" [ 1022.345065] env[62521]: _type = "Task" [ 1022.345065] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.353947] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319027, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.409484] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.423253] env[62521]: DEBUG oslo_concurrency.lockutils [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.619277] env[62521]: INFO nova.compute.manager [-] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Took 1.26 seconds to deallocate network for instance. [ 1022.642502] env[62521]: DEBUG oslo_concurrency.lockutils [req-7064b1a3-75e8-4fde-87b5-a99a71b0eff2 req-ed564d54-a45c-41b5-8eae-a9d8f7cbc214 service nova] Releasing lock "refresh_cache-83fe58c9-920b-422d-be08-e5d53bf551d3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.806268] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319026, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065522} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.806567] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1022.807324] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf8f0ee-e9de-44ee-8a13-5c2032aae706 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.832060] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 83fe58c9-920b-422d-be08-e5d53bf551d3/83fe58c9-920b-422d-be08-e5d53bf551d3.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.832405] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30f77b33-7d42-47bd-a829-cae7ba091a41 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.855188] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319027, 'name': CreateVM_Task, 'duration_secs': 0.375162} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.856242] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1022.856577] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for the task: (returnval){ [ 1022.856577] env[62521]: value = "task-1319028" [ 1022.856577] env[62521]: _type = "Task" [ 1022.856577] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.857182] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.857347] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.857770] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1022.858067] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5dbfa58a-0dbe-4d41-90ae-3d575cbf1d5c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.862708] env[62521]: DEBUG nova.network.neutron [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance_info_cache with network_info: [{"id": "9e954d34-4424-459c-a022-fc039668c616", "address": "fa:16:3e:90:66:a0", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e954d34-44", "ovs_interfaceid": "9e954d34-4424-459c-a022-fc039668c616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.871083] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1022.871083] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522703a9-230b-6690-80c8-23fcf49835b2" [ 1022.871083] env[62521]: _type = "Task" [ 1022.871083] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.871312] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319028, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.883403] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522703a9-230b-6690-80c8-23fcf49835b2, 'name': SearchDatastore_Task, 'duration_secs': 0.012984} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.883693] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.883923] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1022.884171] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.884326] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.884511] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1022.884760] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad38a7ba-a49b-45fd-ba76-cca70efa64f2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.891482] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1022.891681] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1022.892340] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9240fa0f-4a91-41bd-896a-5810f06ccbbe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.896983] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1022.896983] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52df1401-0dd0-ae55-356a-cc69290d194d" [ 1022.896983] env[62521]: _type = "Task" [ 1022.896983] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.906837] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52df1401-0dd0-ae55-356a-cc69290d194d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.909857] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.956440] env[62521]: DEBUG nova.compute.manager [req-2ba77c27-360e-470d-a4b7-61c46fecf85d req-af4a4079-9ca8-40c0-9e4a-9a2a9e097f68 service nova] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Received event network-vif-deleted-b0a026a3-abe7-4087-be49-669d7f6ba99e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.043917] env[62521]: DEBUG nova.network.neutron [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updated VIF entry in instance network info cache for port a2f01aec-5d56-42ef-ac6c-6364f1571174. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1023.044320] env[62521]: DEBUG nova.network.neutron [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating instance_info_cache with network_info: [{"id": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "address": "fa:16:3e:dd:37:93", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2f01aec-5d", "ovs_interfaceid": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.129747] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.130063] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "8b4db43f-bc21-4d40-9410-f643971e53fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.130304] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.194998] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ec9bff-3125-4609-863c-216c3f50cdce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.202873] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245c5e56-b0c6-4587-bf76-01ffd4c77433 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.234201] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8dc9d2c-54b5-42ae-a113-c846e39f5ea2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.241885] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5247d949-05dc-416e-a8d0-73d44c00d4e5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.255186] env[62521]: DEBUG nova.compute.provider_tree [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.365472] env[62521]: DEBUG oslo_concurrency.lockutils [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.371809] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319028, 'name': ReconfigVM_Task, 'duration_secs': 0.305067} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.372260] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 83fe58c9-920b-422d-be08-e5d53bf551d3/83fe58c9-920b-422d-be08-e5d53bf551d3.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.372840] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ad182f2-7076-4d22-a3bf-56102631d20f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.379056] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for the task: (returnval){ [ 1023.379056] env[62521]: value = "task-1319029" [ 1023.379056] env[62521]: _type = "Task" [ 1023.379056] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.386229] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319029, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.416014] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52df1401-0dd0-ae55-356a-cc69290d194d, 'name': SearchDatastore_Task, 'duration_secs': 0.009053} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.420353] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.420567] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86c26c96-039a-4770-bb5a-b99b050a39dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.425453] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1023.425453] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fbbb0f-89a7-ea37-7af3-de6074723959" [ 1023.425453] env[62521]: _type = "Task" [ 1023.425453] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.432632] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fbbb0f-89a7-ea37-7af3-de6074723959, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.546725] env[62521]: DEBUG oslo_concurrency.lockutils [req-45154f1d-6f2b-4ba2-a55a-0ddae3373999 req-99d49e14-101f-443a-81f1-430f58743501 service nova] Releasing lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.633014] env[62521]: DEBUG nova.compute.manager [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1023.758209] env[62521]: DEBUG nova.scheduler.client.report [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.888529] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319029, 'name': Rename_Task, 'duration_secs': 0.142721} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.889513] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1023.890383] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7543fade-7297-4caa-899c-3c4cc0abf7ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.893950] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3619cf7-29b5-46bf-a212-d1763b9cb734 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.915245] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf38446-bdd4-4cf2-942d-694942a7d7f6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.917777] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for the task: (returnval){ [ 1023.917777] env[62521]: value = "task-1319030" [ 1023.917777] env[62521]: _type = "Task" [ 1023.917777] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.926073] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.926498] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance '11e61e72-a311-4c43-bf53-df82ad9d70f7' progress to 83 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1023.937527] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319030, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.943113] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52fbbb0f-89a7-ea37-7af3-de6074723959, 'name': SearchDatastore_Task, 'duration_secs': 0.008784} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.943394] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.943681] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7/6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1023.943944] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d48d9ad-fba3-4bd6-ac62-4373fd366105 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.949694] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1023.949694] env[62521]: value = "task-1319031" [ 1023.949694] env[62521]: _type = "Task" [ 1023.949694] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.957283] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319031, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.153140] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.263541] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.264453] env[62521]: DEBUG nova.compute.manager [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1024.268595] env[62521]: DEBUG oslo_concurrency.lockutils [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.845s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.268889] env[62521]: DEBUG nova.objects.instance [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lazy-loading 'resources' on Instance uuid 31f926a8-777d-4882-afd6-1e4dba3f4e11 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.422301] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.431148] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319030, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.436385] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-00910cef-daa7-481a-996d-78eb6b14e05f tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance '11e61e72-a311-4c43-bf53-df82ad9d70f7' progress to 100 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1024.459534] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319031, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504848} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.459793] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7/6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1024.460044] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1024.460349] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f38c18ca-3253-40c0-b147-4a229ac09cbf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.466477] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1024.466477] env[62521]: value = "task-1319032" [ 1024.466477] env[62521]: _type = "Task" [ 1024.466477] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.475077] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319032, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.772492] env[62521]: DEBUG nova.compute.utils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1024.776978] env[62521]: DEBUG nova.compute.manager [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1024.777165] env[62521]: DEBUG nova.network.neutron [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1024.816729] env[62521]: DEBUG nova.policy [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76d85009c2c1416b9ea049003e6d96a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c533bd4917b4466e9c5c76a9caf8e807', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1024.922890] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.933198] env[62521]: DEBUG oslo_vmware.api [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319030, 'name': PowerOnVM_Task, 'duration_secs': 0.71869} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.933601] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1024.933912] env[62521]: INFO nova.compute.manager [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Took 8.73 seconds to spawn the instance on the hypervisor. [ 1024.934181] env[62521]: DEBUG nova.compute.manager [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.934993] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2069c1c-e614-4eae-9e68-3a7c89a59188 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.939245] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c67e73-685c-4f60-9de7-75dbc94899ec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.953009] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8501d87-e699-40c0-9653-03466e135dcd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.988341] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc29dfea-6351-45d3-806b-37149127708e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.996590] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319032, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064919} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.998834] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1024.999688] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97296ec2-bec5-492f-9308-b62c9f487170 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.003283] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c31057-fefe-4f3b-9dec-c5851616f779 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.027317] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7/6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.036656] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa0cb6bb-a9e8-4299-9871-599409ee04fb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.051078] env[62521]: DEBUG nova.compute.provider_tree [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.058398] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1025.058398] env[62521]: value = "task-1319033" [ 1025.058398] env[62521]: _type = "Task" [ 1025.058398] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.067475] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319033, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.114306] env[62521]: DEBUG nova.network.neutron [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Successfully created port: f8d5ced2-d3b2-4f03-99bc-960baebb3e2a {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.277450] env[62521]: DEBUG nova.compute.manager [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1025.422884] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.460959] env[62521]: INFO nova.compute.manager [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Took 13.97 seconds to build instance. [ 1025.554245] env[62521]: DEBUG nova.scheduler.client.report [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.567798] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319033, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.924602] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.969211] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3479e511-8d1d-46f3-963d-7e9ab08e1d04 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lock "83fe58c9-920b-422d-be08-e5d53bf551d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.484s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.058926] env[62521]: DEBUG oslo_concurrency.lockutils [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.790s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.062433] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.932s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.062433] env[62521]: DEBUG nova.objects.instance [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'resources' on Instance uuid aa266be6-2bf1-445f-a968-2c144d30ecfa {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.074266] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319033, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.084162] env[62521]: INFO nova.scheduler.client.report [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Deleted allocations for instance 31f926a8-777d-4882-afd6-1e4dba3f4e11 [ 1026.286974] env[62521]: DEBUG nova.compute.manager [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1026.311209] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.311468] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.311655] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.311866] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.312031] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.312196] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.312413] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.312573] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.312751] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.312923] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.313110] env[62521]: DEBUG nova.virt.hardware [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.313965] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f39e003-4f96-429b-8d32-aba578895700 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.322537] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b2c49f-5db3-4b8d-8532-35b553b1659b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.423209] env[62521]: DEBUG oslo_vmware.api [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319022, 'name': ReconfigVM_Task, 'duration_secs': 5.757966} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.423451] env[62521]: DEBUG oslo_concurrency.lockutils [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.424155] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Reconfigured VM to detach interface {{(pid=62521) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1026.501868] env[62521]: DEBUG nova.compute.manager [req-45ebe422-8d49-44db-b4a7-676ad9c24749 req-ba2f5fc6-3f14-49e8-9089-47d2e4a86641 service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Received event network-vif-plugged-f8d5ced2-d3b2-4f03-99bc-960baebb3e2a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.502165] env[62521]: DEBUG oslo_concurrency.lockutils [req-45ebe422-8d49-44db-b4a7-676ad9c24749 req-ba2f5fc6-3f14-49e8-9089-47d2e4a86641 service nova] Acquiring lock "60c23400-7eb2-4049-8a3b-77599098e334-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.502380] env[62521]: DEBUG oslo_concurrency.lockutils [req-45ebe422-8d49-44db-b4a7-676ad9c24749 req-ba2f5fc6-3f14-49e8-9089-47d2e4a86641 service nova] Lock "60c23400-7eb2-4049-8a3b-77599098e334-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.502561] env[62521]: DEBUG oslo_concurrency.lockutils [req-45ebe422-8d49-44db-b4a7-676ad9c24749 req-ba2f5fc6-3f14-49e8-9089-47d2e4a86641 service nova] Lock "60c23400-7eb2-4049-8a3b-77599098e334-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.502732] env[62521]: DEBUG nova.compute.manager [req-45ebe422-8d49-44db-b4a7-676ad9c24749 req-ba2f5fc6-3f14-49e8-9089-47d2e4a86641 service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] No waiting events found dispatching network-vif-plugged-f8d5ced2-d3b2-4f03-99bc-960baebb3e2a {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1026.502905] env[62521]: WARNING nova.compute.manager [req-45ebe422-8d49-44db-b4a7-676ad9c24749 req-ba2f5fc6-3f14-49e8-9089-47d2e4a86641 service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Received unexpected event network-vif-plugged-f8d5ced2-d3b2-4f03-99bc-960baebb3e2a for instance with vm_state building and task_state spawning. [ 1026.559421] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquiring lock "83fe58c9-920b-422d-be08-e5d53bf551d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.559658] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lock "83fe58c9-920b-422d-be08-e5d53bf551d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.559887] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquiring lock "83fe58c9-920b-422d-be08-e5d53bf551d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.560086] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lock "83fe58c9-920b-422d-be08-e5d53bf551d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.560265] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lock "83fe58c9-920b-422d-be08-e5d53bf551d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.563057] env[62521]: INFO nova.compute.manager [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Terminating instance [ 1026.568552] env[62521]: DEBUG nova.compute.manager [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1026.568552] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1026.572420] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ef05d5-d139-4850-8f85-eae23b178b77 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.580599] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319033, 'name': ReconfigVM_Task, 'duration_secs': 1.106759} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.582691] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7/6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.583295] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1026.586045] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-18016d06-7e1d-4606-96d2-85dd3b7e248e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.587891] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34936c0b-cbf5-409b-9752-8a5371ec906a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.590037] env[62521]: DEBUG nova.network.neutron [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Successfully updated port: f8d5ced2-d3b2-4f03-99bc-960baebb3e2a {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1026.595701] env[62521]: DEBUG oslo_concurrency.lockutils [None req-775877bb-ffac-45e0-a59b-ae556b742ebf tempest-ServerRescueTestJSONUnderV235-584458748 tempest-ServerRescueTestJSONUnderV235-584458748-project-member] Lock "31f926a8-777d-4882-afd6-1e4dba3f4e11" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.049s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.600409] env[62521]: DEBUG oslo_vmware.api [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for the task: (returnval){ [ 1026.600409] env[62521]: value = "task-1319035" [ 1026.600409] env[62521]: _type = "Task" [ 1026.600409] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.601921] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1026.601921] env[62521]: value = "task-1319034" [ 1026.601921] env[62521]: _type = "Task" [ 1026.601921] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.618629] env[62521]: DEBUG oslo_vmware.api [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319035, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.622805] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319034, 'name': Rename_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.667673] env[62521]: DEBUG nova.compute.manager [req-f7eb9a48-6ff9-4385-88e2-66dc3931cd6e req-8fa814a7-1818-442f-9e3d-992b5891e178 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Received event network-vif-deleted-f5f1a879-9bd4-43e6-815a-307ae9cff9e7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.667896] env[62521]: INFO nova.compute.manager [req-f7eb9a48-6ff9-4385-88e2-66dc3931cd6e req-8fa814a7-1818-442f-9e3d-992b5891e178 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Neutron deleted interface f5f1a879-9bd4-43e6-815a-307ae9cff9e7; detaching it from the instance and deleting it from the info cache [ 1026.668174] env[62521]: DEBUG nova.network.neutron [req-f7eb9a48-6ff9-4385-88e2-66dc3931cd6e req-8fa814a7-1818-442f-9e3d-992b5891e178 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updating instance_info_cache with network_info: [{"id": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "address": "fa:16:3e:37:05:5a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b0cfbc-ec", "ovs_interfaceid": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.750574] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e623017f-ee35-4f92-863b-c9d30131de35 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.757968] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5588ca0-c61d-488e-bbb7-b91db741b62a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.789847] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c997273d-1989-4293-a9e9-628a7db85cf8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.797526] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9751b72d-fccd-490b-aae9-4b47f7a328ee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.810948] env[62521]: DEBUG nova.compute.provider_tree [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.833401] env[62521]: DEBUG oslo_concurrency.lockutils [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.833401] env[62521]: DEBUG oslo_concurrency.lockutils [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.833401] env[62521]: DEBUG nova.compute.manager [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Going to confirm migration 5 {{(pid=62521) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1027.002011] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.002334] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.095413] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "refresh_cache-60c23400-7eb2-4049-8a3b-77599098e334" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.095537] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "refresh_cache-60c23400-7eb2-4049-8a3b-77599098e334" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.095811] env[62521]: DEBUG nova.network.neutron [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.113624] env[62521]: DEBUG oslo_vmware.api [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319035, 'name': PowerOffVM_Task, 'duration_secs': 0.176383} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.114271] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1027.114454] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1027.114727] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eac3cd1d-2986-4b04-8f19-192d533796df {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.119318] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319034, 'name': Rename_Task, 'duration_secs': 0.136965} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.119983] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1027.120252] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca2464ef-80c0-4070-ab10-7d6e4ef61c6c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.126057] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1027.126057] env[62521]: value = "task-1319037" [ 1027.126057] env[62521]: _type = "Task" [ 1027.126057] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.171573] env[62521]: DEBUG oslo_concurrency.lockutils [req-f7eb9a48-6ff9-4385-88e2-66dc3931cd6e req-8fa814a7-1818-442f-9e3d-992b5891e178 service nova] Acquiring lock "b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.171941] env[62521]: DEBUG oslo_concurrency.lockutils [req-f7eb9a48-6ff9-4385-88e2-66dc3931cd6e req-8fa814a7-1818-442f-9e3d-992b5891e178 service nova] Acquired lock "b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.174009] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8bb0b1-1dc6-4d23-9220-8af23de83358 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.193913] env[62521]: DEBUG oslo_concurrency.lockutils [req-f7eb9a48-6ff9-4385-88e2-66dc3931cd6e req-8fa814a7-1818-442f-9e3d-992b5891e178 service nova] Releasing lock "b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.194336] env[62521]: WARNING nova.compute.manager [req-f7eb9a48-6ff9-4385-88e2-66dc3931cd6e req-8fa814a7-1818-442f-9e3d-992b5891e178 service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Detach interface failed, port_id=f5f1a879-9bd4-43e6-815a-307ae9cff9e7, reason: No device with interface-id f5f1a879-9bd4-43e6-815a-307ae9cff9e7 exists on VM: nova.exception.NotFound: No device with interface-id f5f1a879-9bd4-43e6-815a-307ae9cff9e7 exists on VM [ 1027.196161] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1027.196370] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1027.196563] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Deleting the datastore file [datastore1] 83fe58c9-920b-422d-be08-e5d53bf551d3 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.196832] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d95f121d-e185-4b7d-8cd3-8e7839c5af40 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.202773] env[62521]: DEBUG oslo_vmware.api [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for the task: (returnval){ [ 1027.202773] env[62521]: value = "task-1319038" [ 1027.202773] env[62521]: _type = "Task" [ 1027.202773] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.210790] env[62521]: DEBUG oslo_vmware.api [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319038, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.313954] env[62521]: DEBUG nova.scheduler.client.report [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.373819] env[62521]: DEBUG oslo_concurrency.lockutils [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.374249] env[62521]: DEBUG oslo_concurrency.lockutils [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.374508] env[62521]: DEBUG nova.network.neutron [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.374741] env[62521]: DEBUG nova.objects.instance [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'info_cache' on Instance uuid 11e61e72-a311-4c43-bf53-df82ad9d70f7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.505659] env[62521]: DEBUG nova.compute.utils [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1027.635745] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319037, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.641377] env[62521]: DEBUG nova.network.neutron [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1027.646020] env[62521]: DEBUG oslo_concurrency.lockutils [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.646259] env[62521]: DEBUG oslo_concurrency.lockutils [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.646495] env[62521]: DEBUG nova.network.neutron [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.712649] env[62521]: DEBUG oslo_vmware.api [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Task: {'id': task-1319038, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193588} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.712913] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.747726] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1027.748036] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1027.748252] env[62521]: INFO nova.compute.manager [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1027.748545] env[62521]: DEBUG oslo.service.loopingcall [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.752132] env[62521]: DEBUG nova.compute.manager [-] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1027.752243] env[62521]: DEBUG nova.network.neutron [-] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1027.818912] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.757s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.822173] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.668s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.822799] env[62521]: INFO nova.compute.claims [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1027.833363] env[62521]: DEBUG nova.network.neutron [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Updating instance_info_cache with network_info: [{"id": "f8d5ced2-d3b2-4f03-99bc-960baebb3e2a", "address": "fa:16:3e:65:98:33", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8d5ced2-d3", "ovs_interfaceid": "f8d5ced2-d3b2-4f03-99bc-960baebb3e2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.843943] env[62521]: INFO nova.scheduler.client.report [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleted allocations for instance aa266be6-2bf1-445f-a968-2c144d30ecfa [ 1027.919641] env[62521]: DEBUG oslo_concurrency.lockutils [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "b8158051-98b6-47d8-84c5-d1d56515afe8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.919977] env[62521]: DEBUG oslo_concurrency.lockutils [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "b8158051-98b6-47d8-84c5-d1d56515afe8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.920227] env[62521]: DEBUG oslo_concurrency.lockutils [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "b8158051-98b6-47d8-84c5-d1d56515afe8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.920424] env[62521]: DEBUG oslo_concurrency.lockutils [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "b8158051-98b6-47d8-84c5-d1d56515afe8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.920601] env[62521]: DEBUG oslo_concurrency.lockutils [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "b8158051-98b6-47d8-84c5-d1d56515afe8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.922835] env[62521]: INFO nova.compute.manager [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Terminating instance [ 1027.924875] env[62521]: DEBUG nova.compute.manager [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1027.925189] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1027.926047] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7ecebb-0e97-4519-8387-598d8a562ee3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.934415] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1027.934788] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d21819ab-a530-4670-b522-2c28d906ba06 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.941584] env[62521]: DEBUG oslo_vmware.api [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1027.941584] env[62521]: value = "task-1319039" [ 1027.941584] env[62521]: _type = "Task" [ 1027.941584] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.949353] env[62521]: DEBUG oslo_vmware.api [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319039, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.009049] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.139737] env[62521]: DEBUG oslo_vmware.api [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319037, 'name': PowerOnVM_Task, 'duration_secs': 0.539869} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.140474] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1028.140792] env[62521]: INFO nova.compute.manager [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Took 8.73 seconds to spawn the instance on the hypervisor. [ 1028.141203] env[62521]: DEBUG nova.compute.manager [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.142535] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7bf1fef-fc47-4406-bc3a-d801c8b40058 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.335503] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "refresh_cache-60c23400-7eb2-4049-8a3b-77599098e334" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.335865] env[62521]: DEBUG nova.compute.manager [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Instance network_info: |[{"id": "f8d5ced2-d3b2-4f03-99bc-960baebb3e2a", "address": "fa:16:3e:65:98:33", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8d5ced2-d3", "ovs_interfaceid": "f8d5ced2-d3b2-4f03-99bc-960baebb3e2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1028.336493] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:98:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '84aee122-f630-43c5-9cc1-3a38d3819c82', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8d5ced2-d3b2-4f03-99bc-960baebb3e2a', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.344062] env[62521]: DEBUG oslo.service.loopingcall [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.344835] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1028.345083] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d6d31ec-a221-4c5a-9663-0a7cb63805c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.364557] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f868384d-c918-43f5-8659-def9929bb779 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "aa266be6-2bf1-445f-a968-2c144d30ecfa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.176s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.372383] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.372383] env[62521]: value = "task-1319040" [ 1028.372383] env[62521]: _type = "Task" [ 1028.372383] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.379762] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319040, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.406111] env[62521]: DEBUG nova.network.neutron [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updating instance_info_cache with network_info: [{"id": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "address": "fa:16:3e:37:05:5a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b0cfbc-ec", "ovs_interfaceid": "23b0cfbc-ec82-4a35-a248-249e86ecec84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.453835] env[62521]: DEBUG oslo_vmware.api [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319039, 'name': PowerOffVM_Task, 'duration_secs': 0.192205} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.453835] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1028.453835] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1028.453835] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7fbbc90d-92d3-4d02-9559-009e2dc518c0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.521324] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1028.521597] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1028.521766] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleting the datastore file [datastore2] b8158051-98b6-47d8-84c5-d1d56515afe8 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.522098] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6046a3c8-0bcc-430f-8aef-e3309744ff63 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.528658] env[62521]: DEBUG oslo_vmware.api [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1028.528658] env[62521]: value = "task-1319042" [ 1028.528658] env[62521]: _type = "Task" [ 1028.528658] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.536547] env[62521]: DEBUG oslo_vmware.api [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.539442] env[62521]: DEBUG nova.compute.manager [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Received event network-changed-f8d5ced2-d3b2-4f03-99bc-960baebb3e2a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.539678] env[62521]: DEBUG nova.compute.manager [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Refreshing instance network info cache due to event network-changed-f8d5ced2-d3b2-4f03-99bc-960baebb3e2a. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1028.539990] env[62521]: DEBUG oslo_concurrency.lockutils [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] Acquiring lock "refresh_cache-60c23400-7eb2-4049-8a3b-77599098e334" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.540179] env[62521]: DEBUG oslo_concurrency.lockutils [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] Acquired lock "refresh_cache-60c23400-7eb2-4049-8a3b-77599098e334" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.540366] env[62521]: DEBUG nova.network.neutron [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Refreshing network info cache for port f8d5ced2-d3b2-4f03-99bc-960baebb3e2a {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.670102] env[62521]: INFO nova.compute.manager [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Took 17.18 seconds to build instance. [ 1028.728425] env[62521]: DEBUG nova.network.neutron [-] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.785615] env[62521]: DEBUG nova.network.neutron [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance_info_cache with network_info: [{"id": "9e954d34-4424-459c-a022-fc039668c616", "address": "fa:16:3e:90:66:a0", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e954d34-44", "ovs_interfaceid": "9e954d34-4424-459c-a022-fc039668c616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.885035] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319040, 'name': CreateVM_Task, 'duration_secs': 0.335058} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.887625] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.888471] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.888661] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.889017] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1028.889279] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbcce141-dabb-4c5c-8e80-a535444eaed6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.894136] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1028.894136] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f6908c-3be5-b015-d9f3-1b6ca58bd686" [ 1028.894136] env[62521]: _type = "Task" [ 1028.894136] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.901696] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f6908c-3be5-b015-d9f3-1b6ca58bd686, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.908835] env[62521]: DEBUG oslo_concurrency.lockutils [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-b8158051-98b6-47d8-84c5-d1d56515afe8" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.997565] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4419da-7187-4fe5-ae79-9f6345458037 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.005731] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ba401e-89b2-49f2-b15f-2c0090637135 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.038927] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62086d2a-c439-47d1-b1ab-29f8b8af747c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.048797] env[62521]: DEBUG oslo_vmware.api [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319042, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.302484} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.051502] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.051502] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1029.051502] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1029.051502] env[62521]: INFO nova.compute.manager [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1029.051862] env[62521]: DEBUG oslo.service.loopingcall [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.053111] env[62521]: DEBUG nova.compute.manager [-] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.053111] env[62521]: DEBUG nova.network.neutron [-] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1029.056485] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb4b3d3-fce2-4891-b4d7-f13947ba47f7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.070388] env[62521]: DEBUG nova.compute.provider_tree [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1029.096493] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.096770] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.097015] env[62521]: INFO nova.compute.manager [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Attaching volume e62e324b-ba8a-4465-aa3f-4d13a1020baa to /dev/sdb [ 1029.141846] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9f90e3-9d05-4b73-b5d0-41f81f500e06 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.150050] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fecaf7-9c96-4cbb-a554-b7c0c1cc118a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.167050] env[62521]: DEBUG nova.virt.block_device [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Updating existing volume attachment record: 0dd7dc64-327c-4e6d-9549-7a727166255f {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1029.172708] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0efa0f1e-667d-4f9b-aad3-1fd1fdae2879 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.689s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.231472] env[62521]: INFO nova.compute.manager [-] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Took 1.48 seconds to deallocate network for instance. [ 1029.295088] env[62521]: DEBUG oslo_concurrency.lockutils [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.295391] env[62521]: DEBUG nova.objects.instance [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'migration_context' on Instance uuid 11e61e72-a311-4c43-bf53-df82ad9d70f7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.405301] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f6908c-3be5-b015-d9f3-1b6ca58bd686, 'name': SearchDatastore_Task, 'duration_secs': 0.011602} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.405648] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.405886] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.406140] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.406292] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.406474] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.406749] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-588cd59c-f9a6-4824-9184-3a3cc26abe94 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.412863] env[62521]: DEBUG nova.network.neutron [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Updated VIF entry in instance network info cache for port f8d5ced2-d3b2-4f03-99bc-960baebb3e2a. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1029.413462] env[62521]: DEBUG nova.network.neutron [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Updating instance_info_cache with network_info: [{"id": "f8d5ced2-d3b2-4f03-99bc-960baebb3e2a", "address": "fa:16:3e:65:98:33", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8d5ced2-d3", "ovs_interfaceid": "f8d5ced2-d3b2-4f03-99bc-960baebb3e2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.414995] env[62521]: DEBUG oslo_concurrency.lockutils [None req-354b6cf3-429b-4572-b75e-242f902c5ed8 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-b8158051-98b6-47d8-84c5-d1d56515afe8-f5f1a879-9bd4-43e6-815a-307ae9cff9e7" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.608s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.417689] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.417883] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.418834] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-314ac3f8-4c62-455b-8f50-990506d8df2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.425027] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1029.425027] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bc280c-0e5e-3c41-2fcc-74aefdbc55a5" [ 1029.425027] env[62521]: _type = "Task" [ 1029.425027] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.434255] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bc280c-0e5e-3c41-2fcc-74aefdbc55a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.596638] env[62521]: ERROR nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [req-e540cca4-0237-4d03-8216-4b02bb50d320] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e540cca4-0237-4d03-8216-4b02bb50d320"}]} [ 1029.610658] env[62521]: DEBUG nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1029.626101] env[62521]: DEBUG nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1029.626101] env[62521]: DEBUG nova.compute.provider_tree [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1029.643988] env[62521]: DEBUG nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1029.682348] env[62521]: DEBUG nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1029.738744] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.806271] env[62521]: DEBUG nova.objects.base [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Object Instance<11e61e72-a311-4c43-bf53-df82ad9d70f7> lazy-loaded attributes: info_cache,migration_context {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1029.807331] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4578f5f-4ca1-4660-b543-6297607d0ec2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.831582] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98fed686-116c-4b2e-8cf5-d72b872144ee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.834902] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d94436bc-9f40-47b8-a868-b18f29e46408 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.841479] env[62521]: DEBUG oslo_vmware.api [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1029.841479] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e9e986-b1cd-06e0-8451-f90cc1a9a38c" [ 1029.841479] env[62521]: _type = "Task" [ 1029.841479] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.847312] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5cc755-80db-4647-a8ba-007db6150427 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.857215] env[62521]: DEBUG oslo_vmware.api [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e9e986-b1cd-06e0-8451-f90cc1a9a38c, 'name': SearchDatastore_Task, 'duration_secs': 0.009684} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.887105] env[62521]: DEBUG oslo_concurrency.lockutils [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.887105] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96b0563-a898-4e8d-8436-dc3a8f0ba82d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.898360] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09974246-d3ab-40ab-8cd1-e0727768b261 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.907537] env[62521]: DEBUG nova.compute.provider_tree [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1029.918062] env[62521]: DEBUG oslo_concurrency.lockutils [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] Releasing lock "refresh_cache-60c23400-7eb2-4049-8a3b-77599098e334" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.918363] env[62521]: DEBUG nova.compute.manager [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Received event network-vif-deleted-b7d04538-d173-4606-83a1-3c5b32ee6357 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.918569] env[62521]: INFO nova.compute.manager [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Neutron deleted interface b7d04538-d173-4606-83a1-3c5b32ee6357; detaching it from the instance and deleting it from the info cache [ 1029.918745] env[62521]: DEBUG nova.network.neutron [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.937237] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52bc280c-0e5e-3c41-2fcc-74aefdbc55a5, 'name': SearchDatastore_Task, 'duration_secs': 0.010344} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.938141] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b688448c-691c-4ec7-9e4f-634f40d40fb1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.943685] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1029.943685] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5292e03d-7806-6ad6-c376-886c4247f390" [ 1029.943685] env[62521]: _type = "Task" [ 1029.943685] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.953832] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5292e03d-7806-6ad6-c376-886c4247f390, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.160026] env[62521]: DEBUG nova.compute.manager [req-880d8f54-ffbb-4f06-8733-1c1b115885c9 req-efb38907-674e-49a0-8ef9-333186cb423c service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Received event network-vif-deleted-23b0cfbc-ec82-4a35-a248-249e86ecec84 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.160270] env[62521]: INFO nova.compute.manager [req-880d8f54-ffbb-4f06-8733-1c1b115885c9 req-efb38907-674e-49a0-8ef9-333186cb423c service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Neutron deleted interface 23b0cfbc-ec82-4a35-a248-249e86ecec84; detaching it from the instance and deleting it from the info cache [ 1030.160467] env[62521]: DEBUG nova.network.neutron [req-880d8f54-ffbb-4f06-8733-1c1b115885c9 req-efb38907-674e-49a0-8ef9-333186cb423c service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.421451] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e0cecb38-349a-4a22-971f-c9e84b00cd42 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.431297] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdad1c16-c271-4a49-a6c2-fed772ba2ce1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.443354] env[62521]: ERROR nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [req-518e75d9-ba80-4c7d-bf20-f5b255a1dcb0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-518e75d9-ba80-4c7d-bf20-f5b255a1dcb0"}]} [ 1030.454434] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5292e03d-7806-6ad6-c376-886c4247f390, 'name': SearchDatastore_Task, 'duration_secs': 0.012925} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.463875] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.464303] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 60c23400-7eb2-4049-8a3b-77599098e334/60c23400-7eb2-4049-8a3b-77599098e334.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.464948] env[62521]: DEBUG nova.compute.manager [req-a02ea861-74bb-4755-8cc0-d4f9aedb02a3 req-42d7373d-7039-4709-b797-9396d326c56a service nova] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Detach interface failed, port_id=b7d04538-d173-4606-83a1-3c5b32ee6357, reason: Instance 83fe58c9-920b-422d-be08-e5d53bf551d3 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1030.466110] env[62521]: DEBUG nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1030.467966] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8614ff7-28cc-4a3c-9831-3803cf8fe36d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.474793] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1030.474793] env[62521]: value = "task-1319044" [ 1030.474793] env[62521]: _type = "Task" [ 1030.474793] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.484660] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.485463] env[62521]: DEBUG nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1030.485717] env[62521]: DEBUG nova.compute.provider_tree [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1030.497906] env[62521]: DEBUG nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1030.524432] env[62521]: DEBUG nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1030.530741] env[62521]: DEBUG nova.compute.manager [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Stashing vm_state: active {{(pid=62521) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1030.632105] env[62521]: DEBUG nova.network.neutron [-] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.664036] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-22749900-6051-4ce5-9982-2a0cbb4d35d3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.677704] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d4dd0b-0ff3-486a-842e-3099d21ffc5a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.709024] env[62521]: DEBUG nova.compute.manager [req-880d8f54-ffbb-4f06-8733-1c1b115885c9 req-efb38907-674e-49a0-8ef9-333186cb423c service nova] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Detach interface failed, port_id=23b0cfbc-ec82-4a35-a248-249e86ecec84, reason: Instance b8158051-98b6-47d8-84c5-d1d56515afe8 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1030.722915] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b134f2a-a73e-48cc-922a-8f858160c94a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.733224] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3426addf-727a-4320-8bc0-876a3a18a7ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.762684] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95873a7-c289-43e8-b5d2-575cd0322d94 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.771530] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bac6f3-9104-4410-adf1-c20c466ba289 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.787794] env[62521]: DEBUG nova.compute.provider_tree [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1030.984200] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319044, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.049421] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.140071] env[62521]: INFO nova.compute.manager [-] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Took 2.09 seconds to deallocate network for instance. [ 1031.323815] env[62521]: DEBUG nova.scheduler.client.report [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 132 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1031.324202] env[62521]: DEBUG nova.compute.provider_tree [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 132 to 133 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1031.324455] env[62521]: DEBUG nova.compute.provider_tree [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1031.484437] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319044, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.982741} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.484754] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 60c23400-7eb2-4049-8a3b-77599098e334/60c23400-7eb2-4049-8a3b-77599098e334.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1031.484986] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1031.485293] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f2a76ce4-97ba-45a8-bc83-feef5fc60f17 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.491291] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1031.491291] env[62521]: value = "task-1319046" [ 1031.491291] env[62521]: _type = "Task" [ 1031.491291] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.498613] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319046, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.649687] env[62521]: DEBUG oslo_concurrency.lockutils [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.805566] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.805869] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.830329] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.009s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.830865] env[62521]: DEBUG nova.compute.manager [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1031.833406] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.095s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.833632] env[62521]: DEBUG nova.objects.instance [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lazy-loading 'resources' on Instance uuid 83fe58c9-920b-422d-be08-e5d53bf551d3 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.001459] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319046, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068712} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.001736] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.002557] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6022a0-95a8-42d5-b502-954ca0e89ea2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.025473] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 60c23400-7eb2-4049-8a3b-77599098e334/60c23400-7eb2-4049-8a3b-77599098e334.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.025764] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c92b73be-ae6d-45e5-9b64-93c1dd058717 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.044937] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1032.044937] env[62521]: value = "task-1319047" [ 1032.044937] env[62521]: _type = "Task" [ 1032.044937] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.052885] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319047, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.184829] env[62521]: DEBUG nova.compute.manager [req-9a18da6a-eeab-43d9-906c-be02d2d9c9e2 req-4049da0c-db46-4cef-a0d8-4f3b374e4356 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received event network-changed-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.185079] env[62521]: DEBUG nova.compute.manager [req-9a18da6a-eeab-43d9-906c-be02d2d9c9e2 req-4049da0c-db46-4cef-a0d8-4f3b374e4356 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Refreshing instance network info cache due to event network-changed-a2f01aec-5d56-42ef-ac6c-6364f1571174. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1032.185315] env[62521]: DEBUG oslo_concurrency.lockutils [req-9a18da6a-eeab-43d9-906c-be02d2d9c9e2 req-4049da0c-db46-4cef-a0d8-4f3b374e4356 service nova] Acquiring lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.185472] env[62521]: DEBUG oslo_concurrency.lockutils [req-9a18da6a-eeab-43d9-906c-be02d2d9c9e2 req-4049da0c-db46-4cef-a0d8-4f3b374e4356 service nova] Acquired lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.185693] env[62521]: DEBUG nova.network.neutron [req-9a18da6a-eeab-43d9-906c-be02d2d9c9e2 req-4049da0c-db46-4cef-a0d8-4f3b374e4356 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Refreshing network info cache for port a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1032.309235] env[62521]: DEBUG nova.compute.utils [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1032.339471] env[62521]: DEBUG nova.compute.utils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1032.341228] env[62521]: DEBUG nova.compute.manager [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1032.341541] env[62521]: DEBUG nova.network.neutron [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1032.380363] env[62521]: DEBUG nova.policy [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3566601952144168b5d400c41026c81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0dfb5f5bd4646ec95b7c6a2d0434175', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1032.487024] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b3a406-d693-446d-9052-a491fcb122cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.494059] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9368bfce-b8e0-4086-bb8e-3ff9c6218dd2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.522635] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7cb2f9-1d2f-4e5f-8e15-9d93b5c2f1cb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.529572] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6698e6d5-fe7d-46bf-97da-9c620ba22f13 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.542093] env[62521]: DEBUG nova.compute.provider_tree [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.553147] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319047, 'name': ReconfigVM_Task, 'duration_secs': 0.299107} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.553426] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 60c23400-7eb2-4049-8a3b-77599098e334/60c23400-7eb2-4049-8a3b-77599098e334.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.554020] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-febab0cc-4338-4749-b5a1-657c22792831 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.560400] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1032.560400] env[62521]: value = "task-1319048" [ 1032.560400] env[62521]: _type = "Task" [ 1032.560400] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.568177] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319048, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.812404] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.844227] env[62521]: DEBUG nova.compute.manager [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1032.865397] env[62521]: DEBUG nova.network.neutron [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Successfully created port: 9ee94c09-d3e5-465a-9926-639300512161 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1032.911355] env[62521]: DEBUG nova.network.neutron [req-9a18da6a-eeab-43d9-906c-be02d2d9c9e2 req-4049da0c-db46-4cef-a0d8-4f3b374e4356 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updated VIF entry in instance network info cache for port a2f01aec-5d56-42ef-ac6c-6364f1571174. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1032.911714] env[62521]: DEBUG nova.network.neutron [req-9a18da6a-eeab-43d9-906c-be02d2d9c9e2 req-4049da0c-db46-4cef-a0d8-4f3b374e4356 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating instance_info_cache with network_info: [{"id": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "address": "fa:16:3e:dd:37:93", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2f01aec-5d", "ovs_interfaceid": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.048023] env[62521]: DEBUG nova.scheduler.client.report [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.071136] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319048, 'name': Rename_Task, 'duration_secs': 0.148375} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.071466] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1033.071740] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-afca4ff2-fab1-41fc-ab3b-b315ea1255f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.079374] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1033.079374] env[62521]: value = "task-1319049" [ 1033.079374] env[62521]: _type = "Task" [ 1033.079374] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.086422] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319049, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.414167] env[62521]: DEBUG oslo_concurrency.lockutils [req-9a18da6a-eeab-43d9-906c-be02d2d9c9e2 req-4049da0c-db46-4cef-a0d8-4f3b374e4356 service nova] Releasing lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.551014] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.717s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.554056] env[62521]: DEBUG oslo_concurrency.lockutils [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.668s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.585826] env[62521]: INFO nova.scheduler.client.report [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Deleted allocations for instance 83fe58c9-920b-422d-be08-e5d53bf551d3 [ 1033.590906] env[62521]: DEBUG oslo_vmware.api [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319049, 'name': PowerOnVM_Task, 'duration_secs': 0.454998} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.593812] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1033.593812] env[62521]: INFO nova.compute.manager [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Took 7.31 seconds to spawn the instance on the hypervisor. [ 1033.593812] env[62521]: DEBUG nova.compute.manager [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.594625] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9611280-c273-4ad0-bcce-d6acb085842b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.738476] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1033.738730] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282175', 'volume_id': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'name': 'volume-e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c', 'attached_at': '', 'detached_at': '', 'volume_id': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'serial': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1033.739615] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f37082-731e-4eed-9301-c26a0b922b79 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.755830] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764cdbc4-fed7-4944-828c-132bd50784c0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.780515] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] volume-e62e324b-ba8a-4465-aa3f-4d13a1020baa/volume-e62e324b-ba8a-4465-aa3f-4d13a1020baa.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.780766] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-752026a8-3e6e-4ea1-a684-7813868dae61 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.800018] env[62521]: DEBUG oslo_vmware.api [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 1033.800018] env[62521]: value = "task-1319050" [ 1033.800018] env[62521]: _type = "Task" [ 1033.800018] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.810292] env[62521]: DEBUG oslo_vmware.api [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.853022] env[62521]: DEBUG nova.compute.manager [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1033.873866] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.874048] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.874468] env[62521]: INFO nova.compute.manager [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Attaching volume c1234b29-f91d-4e08-ab8b-5699cadfc029 to /dev/sdb [ 1033.884365] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.884618] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.884777] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.884986] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.885154] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.885322] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.885549] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.885739] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.885927] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.886126] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.886317] env[62521]: DEBUG nova.virt.hardware [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.887210] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba9ad60-2886-499d-874b-017b8c538261 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.895718] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019613fc-2ae4-44dd-9739-17a0da7cdaf4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.911047] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c55de40-8831-4a55-b534-c770544503c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.917096] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6b550d-60f6-4ca4-8302-7345cdbe81c6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.929679] env[62521]: DEBUG nova.virt.block_device [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Updating existing volume attachment record: 4f02c90c-37ec-40bd-aad9-ab7a79110473 {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1034.097581] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2e3926cd-d4fa-46fe-bc21-57ec47dbe4d0 tempest-ServerMetadataNegativeTestJSON-825824358 tempest-ServerMetadataNegativeTestJSON-825824358-project-member] Lock "83fe58c9-920b-422d-be08-e5d53bf551d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.538s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.119717] env[62521]: INFO nova.compute.manager [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Took 12.42 seconds to build instance. [ 1034.242700] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e67116-6fb9-462f-bec9-a75215ec8ffd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.251860] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a27547c-2c27-4301-b923-69d44d33e62c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.284671] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32708a43-c53b-474b-afed-5965dabdbfc5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.292912] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7985da-db3a-4d05-a729-4b1c0a8914fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.309991] env[62521]: DEBUG nova.compute.provider_tree [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.321960] env[62521]: DEBUG oslo_vmware.api [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319050, 'name': ReconfigVM_Task, 'duration_secs': 0.389468} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.322286] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Reconfigured VM instance instance-00000057 to attach disk [datastore2] volume-e62e324b-ba8a-4465-aa3f-4d13a1020baa/volume-e62e324b-ba8a-4465-aa3f-4d13a1020baa.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.328458] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eec94e50-1dca-444a-ac19-6e23ace4ade5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.345736] env[62521]: DEBUG oslo_vmware.api [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 1034.345736] env[62521]: value = "task-1319052" [ 1034.345736] env[62521]: _type = "Task" [ 1034.345736] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.354979] env[62521]: DEBUG oslo_vmware.api [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319052, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.613185] env[62521]: DEBUG nova.compute.manager [req-5fb55f30-609c-4d0d-9b30-ba9af4109477 req-56fda89c-5a3e-4094-bca6-78f1dd5c0620 service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Received event network-vif-plugged-9ee94c09-d3e5-465a-9926-639300512161 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1034.613578] env[62521]: DEBUG oslo_concurrency.lockutils [req-5fb55f30-609c-4d0d-9b30-ba9af4109477 req-56fda89c-5a3e-4094-bca6-78f1dd5c0620 service nova] Acquiring lock "8b4db43f-bc21-4d40-9410-f643971e53fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.613659] env[62521]: DEBUG oslo_concurrency.lockutils [req-5fb55f30-609c-4d0d-9b30-ba9af4109477 req-56fda89c-5a3e-4094-bca6-78f1dd5c0620 service nova] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.613924] env[62521]: DEBUG oslo_concurrency.lockutils [req-5fb55f30-609c-4d0d-9b30-ba9af4109477 req-56fda89c-5a3e-4094-bca6-78f1dd5c0620 service nova] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.614080] env[62521]: DEBUG nova.compute.manager [req-5fb55f30-609c-4d0d-9b30-ba9af4109477 req-56fda89c-5a3e-4094-bca6-78f1dd5c0620 service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] No waiting events found dispatching network-vif-plugged-9ee94c09-d3e5-465a-9926-639300512161 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1034.614192] env[62521]: WARNING nova.compute.manager [req-5fb55f30-609c-4d0d-9b30-ba9af4109477 req-56fda89c-5a3e-4094-bca6-78f1dd5c0620 service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Received unexpected event network-vif-plugged-9ee94c09-d3e5-465a-9926-639300512161 for instance with vm_state building and task_state spawning. [ 1034.621992] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0f3e99d9-3e9d-46c7-afa1-c9be82bea09e tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "60c23400-7eb2-4049-8a3b-77599098e334" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.927s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.800910] env[62521]: DEBUG nova.network.neutron [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Successfully updated port: 9ee94c09-d3e5-465a-9926-639300512161 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1034.817996] env[62521]: DEBUG nova.scheduler.client.report [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.856456] env[62521]: DEBUG oslo_vmware.api [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319052, 'name': ReconfigVM_Task, 'duration_secs': 0.151893} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.857058] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282175', 'volume_id': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'name': 'volume-e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c', 'attached_at': '', 'detached_at': '', 'volume_id': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'serial': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1035.176767] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "60c23400-7eb2-4049-8a3b-77599098e334" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.176948] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "60c23400-7eb2-4049-8a3b-77599098e334" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.177174] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "60c23400-7eb2-4049-8a3b-77599098e334-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.177428] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "60c23400-7eb2-4049-8a3b-77599098e334-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.177533] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "60c23400-7eb2-4049-8a3b-77599098e334-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.180195] env[62521]: INFO nova.compute.manager [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Terminating instance [ 1035.182168] env[62521]: DEBUG nova.compute.manager [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1035.182974] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.183315] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd0d7fb-9437-472b-898e-183880c32cbf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.191121] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.191420] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83482187-281c-44a6-b36f-6867fe113f39 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.197540] env[62521]: DEBUG oslo_vmware.api [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1035.197540] env[62521]: value = "task-1319053" [ 1035.197540] env[62521]: _type = "Task" [ 1035.197540] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.206692] env[62521]: DEBUG oslo_vmware.api [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.303950] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.304208] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.304421] env[62521]: DEBUG nova.network.neutron [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1035.711195] env[62521]: DEBUG oslo_vmware.api [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319053, 'name': PowerOffVM_Task, 'duration_secs': 0.20754} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.712035] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1035.712035] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1035.712511] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55624c3e-be40-4b3c-885c-787343926e20 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.778320] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1035.778708] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1035.779058] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleting the datastore file [datastore2] 60c23400-7eb2-4049-8a3b-77599098e334 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.779358] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf8557c1-d486-48aa-8cfc-9654d3b9d205 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.785694] env[62521]: DEBUG oslo_vmware.api [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1035.785694] env[62521]: value = "task-1319055" [ 1035.785694] env[62521]: _type = "Task" [ 1035.785694] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.793380] env[62521]: DEBUG oslo_vmware.api [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319055, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.835452] env[62521]: DEBUG oslo_concurrency.lockutils [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.282s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.835547] env[62521]: DEBUG nova.compute.manager [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62521) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1035.838468] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.789s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.850950] env[62521]: DEBUG nova.network.neutron [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1035.904776] env[62521]: DEBUG nova.objects.instance [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'flavor' on Instance uuid 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.060801] env[62521]: DEBUG nova.network.neutron [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Updating instance_info_cache with network_info: [{"id": "9ee94c09-d3e5-465a-9926-639300512161", "address": "fa:16:3e:0b:5a:3b", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ee94c09-d3", "ovs_interfaceid": "9ee94c09-d3e5-465a-9926-639300512161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.295589] env[62521]: DEBUG oslo_vmware.api [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319055, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.241787} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.295949] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.296162] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.296369] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.296617] env[62521]: INFO nova.compute.manager [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1036.296897] env[62521]: DEBUG oslo.service.loopingcall [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.297110] env[62521]: DEBUG nova.compute.manager [-] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.297209] env[62521]: DEBUG nova.network.neutron [-] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1036.346081] env[62521]: INFO nova.compute.claims [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.412010] env[62521]: INFO nova.scheduler.client.report [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleted allocation for migration 1db23eae-760d-44f6-9059-4145c6530e07 [ 1036.413198] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae26c070-7678-453a-b900-8b04936ef1a8 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.316s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.563818] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.564200] env[62521]: DEBUG nova.compute.manager [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Instance network_info: |[{"id": "9ee94c09-d3e5-465a-9926-639300512161", "address": "fa:16:3e:0b:5a:3b", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ee94c09-d3", "ovs_interfaceid": "9ee94c09-d3e5-465a-9926-639300512161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1036.564680] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:5a:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '304be4f7-4e36-4468-9ef4-e457341cef18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9ee94c09-d3e5-465a-9926-639300512161', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1036.572795] env[62521]: DEBUG oslo.service.loopingcall [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.573058] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1036.573312] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c589fcef-ce91-4455-97b9-f433931d313c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.594813] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1036.594813] env[62521]: value = "task-1319057" [ 1036.594813] env[62521]: _type = "Task" [ 1036.594813] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.602469] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319057, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.645640] env[62521]: DEBUG nova.compute.manager [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Received event network-changed-9ee94c09-d3e5-465a-9926-639300512161 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1036.645879] env[62521]: DEBUG nova.compute.manager [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Refreshing instance network info cache due to event network-changed-9ee94c09-d3e5-465a-9926-639300512161. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1036.646154] env[62521]: DEBUG oslo_concurrency.lockutils [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] Acquiring lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.646353] env[62521]: DEBUG oslo_concurrency.lockutils [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] Acquired lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.646574] env[62521]: DEBUG nova.network.neutron [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Refreshing network info cache for port 9ee94c09-d3e5-465a-9926-639300512161 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1036.854738] env[62521]: INFO nova.compute.resource_tracker [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating resource usage from migration ff0bde79-bf70-463f-ae4e-b1a86e2a6db9 [ 1036.918577] env[62521]: DEBUG oslo_concurrency.lockutils [None req-71589819-25ca-4546-b595-ada9489b18fa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.085s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.036329] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f51d5c-bb6e-46d4-8e79-011e971ef744 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.044573] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c2a2fd-307c-428c-ae45-e8c9874e65a9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.084126] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1141c656-dfe3-4f54-a903-261ec9bc53b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.093085] env[62521]: DEBUG nova.network.neutron [-] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.097082] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599b10fb-122c-4370-ac95-288959549158 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.117348] env[62521]: DEBUG nova.compute.provider_tree [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.122897] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319057, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.302215] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.302390] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.376743] env[62521]: DEBUG nova.network.neutron [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Updated VIF entry in instance network info cache for port 9ee94c09-d3e5-465a-9926-639300512161. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1037.377136] env[62521]: DEBUG nova.network.neutron [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Updating instance_info_cache with network_info: [{"id": "9ee94c09-d3e5-465a-9926-639300512161", "address": "fa:16:3e:0b:5a:3b", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ee94c09-d3", "ovs_interfaceid": "9ee94c09-d3e5-465a-9926-639300512161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.406444] env[62521]: DEBUG nova.objects.instance [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'flavor' on Instance uuid 11e61e72-a311-4c43-bf53-df82ad9d70f7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.602022] env[62521]: INFO nova.compute.manager [-] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Took 1.30 seconds to deallocate network for instance. [ 1037.621027] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319057, 'name': CreateVM_Task, 'duration_secs': 0.653603} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.621027] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1037.621027] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.621027] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.621027] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.621027] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b2fc865-428f-4913-9578-8654d9c41013 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.624419] env[62521]: DEBUG nova.scheduler.client.report [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.632703] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1037.632703] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f349ca-115d-a7a6-2cbc-6f55e9289380" [ 1037.632703] env[62521]: _type = "Task" [ 1037.632703] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.648074] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f349ca-115d-a7a6-2cbc-6f55e9289380, 'name': SearchDatastore_Task, 'duration_secs': 0.009239} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.648964] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.649225] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.649461] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.651348] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.651348] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.651348] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a1217b0-12f9-4875-888e-a5955ce4629c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.661867] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.661867] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1037.661867] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af2c4b6b-aaf5-4723-adf0-47636fdb7582 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.666620] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1037.666620] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b9847c-e745-b4a4-b2be-ffc09353c84c" [ 1037.666620] env[62521]: _type = "Task" [ 1037.666620] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.675539] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b9847c-e745-b4a4-b2be-ffc09353c84c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.807469] env[62521]: DEBUG nova.compute.utils [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1037.880594] env[62521]: DEBUG oslo_concurrency.lockutils [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] Releasing lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.880594] env[62521]: DEBUG nova.compute.manager [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Received event network-vif-deleted-f8d5ced2-d3b2-4f03-99bc-960baebb3e2a {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.880594] env[62521]: INFO nova.compute.manager [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Neutron deleted interface f8d5ced2-d3b2-4f03-99bc-960baebb3e2a; detaching it from the instance and deleting it from the info cache [ 1037.880594] env[62521]: DEBUG nova.network.neutron [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.911138] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.911487] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.911487] env[62521]: DEBUG nova.network.neutron [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1037.911664] env[62521]: DEBUG nova.objects.instance [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'info_cache' on Instance uuid 11e61e72-a311-4c43-bf53-df82ad9d70f7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.115371] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.129492] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.291s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.130092] env[62521]: INFO nova.compute.manager [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Migrating [ 1038.137194] env[62521]: DEBUG oslo_concurrency.lockutils [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.487s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.137194] env[62521]: DEBUG nova.objects.instance [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'resources' on Instance uuid b8158051-98b6-47d8-84c5-d1d56515afe8 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.178271] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b9847c-e745-b4a4-b2be-ffc09353c84c, 'name': SearchDatastore_Task, 'duration_secs': 0.009785} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.179507] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77daca68-b364-4b74-a5f3-0ad11223c2b6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.185518] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1038.185518] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5289cc87-0552-3a09-2aff-a5b57d1dde35" [ 1038.185518] env[62521]: _type = "Task" [ 1038.185518] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.193848] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5289cc87-0552-3a09-2aff-a5b57d1dde35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.310040] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.382949] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5091c72-7c57-4580-a661-420615a07525 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.392939] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5c807d-8129-4506-acb0-72c4a25a8ed5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.422478] env[62521]: DEBUG nova.objects.base [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Object Instance<11e61e72-a311-4c43-bf53-df82ad9d70f7> lazy-loaded attributes: flavor,info_cache {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1038.423742] env[62521]: DEBUG nova.compute.manager [req-0adafc0c-bc21-430e-be80-0d025956b46b req-c96e5a98-708a-4e22-8819-4a6ff3cacecb service nova] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Detach interface failed, port_id=f8d5ced2-d3b2-4f03-99bc-960baebb3e2a, reason: Instance 60c23400-7eb2-4049-8a3b-77599098e334 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1038.475923] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1038.476209] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282176', 'volume_id': 'c1234b29-f91d-4e08-ab8b-5699cadfc029', 'name': 'volume-c1234b29-f91d-4e08-ab8b-5699cadfc029', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a68d4cb7-bfad-48f6-8aeb-bca04248a0d7', 'attached_at': '', 'detached_at': '', 'volume_id': 'c1234b29-f91d-4e08-ab8b-5699cadfc029', 'serial': 'c1234b29-f91d-4e08-ab8b-5699cadfc029'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1038.477459] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886eaf01-19cf-4ad7-a495-8ad90e2c6e89 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.495721] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ee5037-e27e-4b82-922b-de2af65c05c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.523919] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] volume-c1234b29-f91d-4e08-ab8b-5699cadfc029/volume-c1234b29-f91d-4e08-ab8b-5699cadfc029.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.524883] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f104ae72-ec0e-4008-8ea3-640a18d5f6d3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.542738] env[62521]: DEBUG oslo_vmware.api [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1038.542738] env[62521]: value = "task-1319058" [ 1038.542738] env[62521]: _type = "Task" [ 1038.542738] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.550689] env[62521]: DEBUG oslo_vmware.api [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319058, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.647179] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.647382] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.647565] env[62521]: DEBUG nova.network.neutron [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1038.695488] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5289cc87-0552-3a09-2aff-a5b57d1dde35, 'name': SearchDatastore_Task, 'duration_secs': 0.013705} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.695788] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.696056] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 8b4db43f-bc21-4d40-9410-f643971e53fc/8b4db43f-bc21-4d40-9410-f643971e53fc.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1038.698449] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de30de8f-f24b-4dce-850f-5c8af2e134b3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.706295] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1038.706295] env[62521]: value = "task-1319059" [ 1038.706295] env[62521]: _type = "Task" [ 1038.706295] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.713913] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319059, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.801573] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c173a0-0df5-4d25-9d90-59b4f2320cde {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.809462] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20553560-70c7-4a6a-ab91-69a08703c156 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.842327] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490458d9-15dc-4bdc-b6a5-71cf2409e748 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.850522] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab0be66-f372-4094-a4b0-522ffeae6306 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.866469] env[62521]: DEBUG nova.compute.provider_tree [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.054585] env[62521]: DEBUG oslo_vmware.api [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319058, 'name': ReconfigVM_Task, 'duration_secs': 0.406011} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.054945] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Reconfigured VM instance instance-00000058 to attach disk [datastore2] volume-c1234b29-f91d-4e08-ab8b-5699cadfc029/volume-c1234b29-f91d-4e08-ab8b-5699cadfc029.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.060659] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-982d3fec-a337-4dda-b039-1d2eed61e327 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.081888] env[62521]: DEBUG oslo_vmware.api [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1039.081888] env[62521]: value = "task-1319060" [ 1039.081888] env[62521]: _type = "Task" [ 1039.081888] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.098667] env[62521]: DEBUG oslo_vmware.api [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319060, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.103366] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.103633] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.221827] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319059, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.324736] env[62521]: DEBUG nova.network.neutron [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance_info_cache with network_info: [{"id": "9e954d34-4424-459c-a022-fc039668c616", "address": "fa:16:3e:90:66:a0", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e954d34-44", "ovs_interfaceid": "9e954d34-4424-459c-a022-fc039668c616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.372484] env[62521]: DEBUG nova.scheduler.client.report [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.383883] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.384165] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.384404] env[62521]: INFO nova.compute.manager [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Attaching volume cfc4d873-e564-4983-a0f5-9f3afecbb1ea to /dev/sdc [ 1039.424710] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d43fb5-5c9e-420d-892b-396dbd6f8a66 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.431730] env[62521]: DEBUG nova.network.neutron [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [{"id": "8824a231-01df-41e5-9bea-9071d8b41cb8", "address": "fa:16:3e:2e:cc:1b", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8824a231-01", "ovs_interfaceid": "8824a231-01df-41e5-9bea-9071d8b41cb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.438019] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0e9e03-e997-4125-8bb5-821f68782cd3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.451575] env[62521]: DEBUG nova.virt.block_device [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Updating existing volume attachment record: 502910b3-782f-4770-807e-5abd69706e1f {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1039.595564] env[62521]: DEBUG oslo_vmware.api [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319060, 'name': ReconfigVM_Task, 'duration_secs': 0.178777} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.595926] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282176', 'volume_id': 'c1234b29-f91d-4e08-ab8b-5699cadfc029', 'name': 'volume-c1234b29-f91d-4e08-ab8b-5699cadfc029', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a68d4cb7-bfad-48f6-8aeb-bca04248a0d7', 'attached_at': '', 'detached_at': '', 'volume_id': 'c1234b29-f91d-4e08-ab8b-5699cadfc029', 'serial': 'c1234b29-f91d-4e08-ab8b-5699cadfc029'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1039.606115] env[62521]: DEBUG nova.compute.manager [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.716986] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319059, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543175} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.717345] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 8b4db43f-bc21-4d40-9410-f643971e53fc/8b4db43f-bc21-4d40-9410-f643971e53fc.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1039.717584] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.718482] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8695b8c7-7411-418e-9b1b-0ae48b2da313 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.726270] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1039.726270] env[62521]: value = "task-1319062" [ 1039.726270] env[62521]: _type = "Task" [ 1039.726270] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.734399] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319062, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.830306] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-11e61e72-a311-4c43-bf53-df82ad9d70f7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.877958] env[62521]: DEBUG oslo_concurrency.lockutils [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.882468] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.766s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.882468] env[62521]: DEBUG nova.objects.instance [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lazy-loading 'resources' on Instance uuid 60c23400-7eb2-4049-8a3b-77599098e334 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.898877] env[62521]: INFO nova.scheduler.client.report [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleted allocations for instance b8158051-98b6-47d8-84c5-d1d56515afe8 [ 1039.934791] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.024303] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.024534] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.133342] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.236432] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319062, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072625} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.236755] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.237559] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a71a725-ff2d-4ac8-ac05-7fbf8d0d9f2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.260937] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 8b4db43f-bc21-4d40-9410-f643971e53fc/8b4db43f-bc21-4d40-9410-f643971e53fc.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.261254] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-398e1c37-325b-4c77-a4cc-7371d25e32b4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.279675] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1040.279675] env[62521]: value = "task-1319063" [ 1040.279675] env[62521]: _type = "Task" [ 1040.279675] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.287524] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319063, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.333948] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1040.334295] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d7a7a35-071d-45fc-a3bb-b1ff58632a05 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.341340] env[62521]: DEBUG oslo_vmware.api [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1040.341340] env[62521]: value = "task-1319064" [ 1040.341340] env[62521]: _type = "Task" [ 1040.341340] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.348746] env[62521]: DEBUG oslo_vmware.api [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319064, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.405902] env[62521]: DEBUG oslo_concurrency.lockutils [None req-89d504fe-280f-47ef-af68-d3cd6e35022f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "b8158051-98b6-47d8-84c5-d1d56515afe8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.486s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.527285] env[62521]: DEBUG nova.compute.manager [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1040.536384] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a66074a-b11d-4533-94b0-5e61f00ba6c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.543700] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39b35a2-496f-4928-8324-96a64ee387cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.572975] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0a5e29-59e8-47f5-9404-0a51c697bf12 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.580659] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f7627b-1772-41cb-8262-9da0a5ede0a9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.593818] env[62521]: DEBUG nova.compute.provider_tree [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.647738] env[62521]: DEBUG nova.objects.instance [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lazy-loading 'flavor' on Instance uuid a68d4cb7-bfad-48f6-8aeb-bca04248a0d7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.791038] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319063, 'name': ReconfigVM_Task, 'duration_secs': 0.255813} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.791038] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 8b4db43f-bc21-4d40-9410-f643971e53fc/8b4db43f-bc21-4d40-9410-f643971e53fc.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.791038] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7258230-a9ac-4e2a-ae61-a717e2c94d8a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.797743] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1040.797743] env[62521]: value = "task-1319065" [ 1040.797743] env[62521]: _type = "Task" [ 1040.797743] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.805695] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319065, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.851224] env[62521]: DEBUG oslo_vmware.api [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319064, 'name': PowerOnVM_Task, 'duration_secs': 0.387126} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.851558] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1040.851773] env[62521]: DEBUG nova.compute.manager [None req-4e9d538a-127a-46a6-aaed-b3575ceb6fa0 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1040.852640] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb51b67-15bd-4aee-b7e3-e3b80459fe67 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.045504] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.096780] env[62521]: DEBUG nova.scheduler.client.report [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.154723] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7f58e008-6b47-431a-8cf7-39c67009bed6 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.281s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.312152] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319065, 'name': Rename_Task, 'duration_secs': 0.150947} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.312624] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1041.312985] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ad2c3e7-919b-4dd4-9a1b-033671322f10 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.320701] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1041.320701] env[62521]: value = "task-1319066" [ 1041.320701] env[62521]: _type = "Task" [ 1041.320701] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.331879] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319066, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.430673] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.430970] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.453508] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0310bb6e-c70e-440d-9af5-6b170b8ef412 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.472426] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance '7b5afeab-f04e-453a-a3ed-09e3a725e8d6' progress to 0 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1041.604013] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.720s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.606870] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.474s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.609106] env[62521]: INFO nova.compute.claims [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1041.638922] env[62521]: INFO nova.scheduler.client.report [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted allocations for instance 60c23400-7eb2-4049-8a3b-77599098e334 [ 1041.830343] env[62521]: DEBUG oslo_vmware.api [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319066, 'name': PowerOnVM_Task, 'duration_secs': 0.478827} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.830648] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1041.830858] env[62521]: INFO nova.compute.manager [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Took 7.98 seconds to spawn the instance on the hypervisor. [ 1041.831055] env[62521]: DEBUG nova.compute.manager [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1041.831823] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0ad7fec-fa9d-4b8c-ac8f-628c402528ea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.934362] env[62521]: INFO nova.compute.manager [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Detaching volume c1234b29-f91d-4e08-ab8b-5699cadfc029 [ 1041.975577] env[62521]: INFO nova.virt.block_device [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Attempting to driver detach volume c1234b29-f91d-4e08-ab8b-5699cadfc029 from mountpoint /dev/sdb [ 1041.975781] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1041.976055] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282176', 'volume_id': 'c1234b29-f91d-4e08-ab8b-5699cadfc029', 'name': 'volume-c1234b29-f91d-4e08-ab8b-5699cadfc029', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a68d4cb7-bfad-48f6-8aeb-bca04248a0d7', 'attached_at': '', 'detached_at': '', 'volume_id': 'c1234b29-f91d-4e08-ab8b-5699cadfc029', 'serial': 'c1234b29-f91d-4e08-ab8b-5699cadfc029'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1041.977673] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1041.978516] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c44164-560c-4eee-bb8b-80c0e9492bb8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.981173] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dbfc964d-6a0c-40b1-8f43-a9f83bc46835 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.002835] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1042.002835] env[62521]: value = "task-1319068" [ 1042.002835] env[62521]: _type = "Task" [ 1042.002835] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.003591] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8637d9b-d74c-431b-b02e-89c1657688ab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.016288] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e1a2af-59b4-477e-aab8-75fb7063a83b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.018802] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319068, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.037180] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2d453a-4e07-4f20-a7e3-0ac8d9791969 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.052295] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] The volume has not been displaced from its original location: [datastore2] volume-c1234b29-f91d-4e08-ab8b-5699cadfc029/volume-c1234b29-f91d-4e08-ab8b-5699cadfc029.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1042.057673] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Reconfiguring VM instance instance-00000058 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1042.059016] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1b942dd-5015-4fa3-af72-ffdbdb6aba38 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.081596] env[62521]: DEBUG oslo_vmware.api [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1042.081596] env[62521]: value = "task-1319069" [ 1042.081596] env[62521]: _type = "Task" [ 1042.081596] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.089960] env[62521]: DEBUG oslo_vmware.api [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319069, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.145907] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7c552718-67cd-4ce5-9ad9-25e99ca8e4eb tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "60c23400-7eb2-4049-8a3b-77599098e334" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.968s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.350421] env[62521]: INFO nova.compute.manager [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Took 18.21 seconds to build instance. [ 1042.516287] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319068, 'name': PowerOffVM_Task, 'duration_secs': 0.258399} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.516562] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1042.516758] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance '7b5afeab-f04e-453a-a3ed-09e3a725e8d6' progress to 17 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1042.577138] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.577899] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.577899] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "11e61e72-a311-4c43-bf53-df82ad9d70f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.577899] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.578099] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.580408] env[62521]: INFO nova.compute.manager [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Terminating instance [ 1042.582398] env[62521]: DEBUG nova.compute.manager [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1042.582616] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1042.583522] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de8ff02-3447-4a8b-a251-c887b879197e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.597838] env[62521]: DEBUG oslo_vmware.api [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319069, 'name': ReconfigVM_Task, 'duration_secs': 0.306263} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.598475] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1042.598475] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Reconfigured VM instance instance-00000058 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1042.603341] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7a34050-3466-475d-9cdf-547055f51bfa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.604889] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45a8a88e-0a50-4a80-afa0-6990e2d9fba9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.624626] env[62521]: DEBUG oslo_vmware.api [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1042.624626] env[62521]: value = "task-1319071" [ 1042.624626] env[62521]: _type = "Task" [ 1042.624626] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.625472] env[62521]: DEBUG oslo_vmware.api [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1042.625472] env[62521]: value = "task-1319070" [ 1042.625472] env[62521]: _type = "Task" [ 1042.625472] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.638897] env[62521]: DEBUG oslo_vmware.api [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.644525] env[62521]: DEBUG oslo_vmware.api [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319070, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.781972] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffbe3986-f4d5-4c71-b534-209086aa989a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.789424] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153a30c5-f2b5-49c9-a79d-c1647c6dd892 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.821354] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717b8867-0536-40c3-9c41-43e078210f14 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.829897] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bbae94a-fc8b-4398-806d-d67779aca8f5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.846888] env[62521]: DEBUG nova.compute.provider_tree [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.852534] env[62521]: DEBUG oslo_concurrency.lockutils [None req-266786bb-774a-48b4-b0bc-01184db63b99 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.722s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.889438] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "bc7d76b1-76b2-49a7-8188-0672312767a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.889684] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "bc7d76b1-76b2-49a7-8188-0672312767a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.022967] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1043.023128] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1043.023229] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.023416] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1043.023569] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.023724] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1043.023965] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1043.024161] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1043.024354] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1043.024527] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1043.024706] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1043.029930] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10508822-6a53-43e9-80d2-28b867a565c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.045344] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1043.045344] env[62521]: value = "task-1319072" [ 1043.045344] env[62521]: _type = "Task" [ 1043.045344] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.053124] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319072, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.139163] env[62521]: DEBUG oslo_vmware.api [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319070, 'name': PowerOffVM_Task, 'duration_secs': 0.182357} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.142551] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1043.142740] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1043.143327] env[62521]: DEBUG oslo_vmware.api [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319071, 'name': ReconfigVM_Task, 'duration_secs': 0.152995} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.143531] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aee2e36d-04e4-4f30-b460-10e5e9d894d3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.145053] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282176', 'volume_id': 'c1234b29-f91d-4e08-ab8b-5699cadfc029', 'name': 'volume-c1234b29-f91d-4e08-ab8b-5699cadfc029', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a68d4cb7-bfad-48f6-8aeb-bca04248a0d7', 'attached_at': '', 'detached_at': '', 'volume_id': 'c1234b29-f91d-4e08-ab8b-5699cadfc029', 'serial': 'c1234b29-f91d-4e08-ab8b-5699cadfc029'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1043.159995] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.160255] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.205985] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1043.206252] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1043.206444] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleting the datastore file [datastore2] 11e61e72-a311-4c43-bf53-df82ad9d70f7 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.206716] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4333b542-d489-4b75-9c8b-e5f127982667 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.212664] env[62521]: DEBUG oslo_vmware.api [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1043.212664] env[62521]: value = "task-1319074" [ 1043.212664] env[62521]: _type = "Task" [ 1043.212664] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.220155] env[62521]: DEBUG oslo_vmware.api [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.339722] env[62521]: DEBUG oslo_concurrency.lockutils [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "8b4db43f-bc21-4d40-9410-f643971e53fc" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.339952] env[62521]: DEBUG oslo_concurrency.lockutils [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.340231] env[62521]: INFO nova.compute.manager [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Shelving [ 1043.349900] env[62521]: DEBUG nova.scheduler.client.report [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.391455] env[62521]: DEBUG nova.compute.manager [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1043.554844] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319072, 'name': ReconfigVM_Task, 'duration_secs': 0.208977} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.555194] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance '7b5afeab-f04e-453a-a3ed-09e3a725e8d6' progress to 33 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1043.662607] env[62521]: DEBUG nova.compute.manager [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1043.687336] env[62521]: DEBUG nova.objects.instance [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lazy-loading 'flavor' on Instance uuid a68d4cb7-bfad-48f6-8aeb-bca04248a0d7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.721956] env[62521]: DEBUG oslo_vmware.api [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.847523] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1043.847759] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-935d6644-0f76-426a-97ea-73b1c3d9833c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.854730] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.248s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.855201] env[62521]: DEBUG nova.compute.manager [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1043.857709] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1043.857709] env[62521]: value = "task-1319075" [ 1043.857709] env[62521]: _type = "Task" [ 1043.857709] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.858140] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.813s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.859506] env[62521]: INFO nova.compute.claims [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1043.870150] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319075, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.911539] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.996969] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1043.997717] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282178', 'volume_id': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'name': 'volume-cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c', 'attached_at': '', 'detached_at': '', 'volume_id': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'serial': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1043.998694] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6d6fb9-aafe-4d36-bf2d-130107f65977 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.015577] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d5344c-ba0a-4325-a9bc-f10c7a721c5a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.043732] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] volume-cfc4d873-e564-4983-a0f5-9f3afecbb1ea/volume-cfc4d873-e564-4983-a0f5-9f3afecbb1ea.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1044.044072] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4c57d1f-f001-421e-8f18-585a1f580431 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.061500] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.061742] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.061904] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.062101] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.062258] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.062410] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.062620] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.062784] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.062975] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.063186] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.063369] env[62521]: DEBUG nova.virt.hardware [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.068442] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Reconfiguring VM instance instance-00000044 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1044.069900] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e05c40a8-020b-4911-9acb-3a9fd9d558e6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.082826] env[62521]: DEBUG oslo_vmware.api [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 1044.082826] env[62521]: value = "task-1319076" [ 1044.082826] env[62521]: _type = "Task" [ 1044.082826] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.089526] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1044.089526] env[62521]: value = "task-1319077" [ 1044.089526] env[62521]: _type = "Task" [ 1044.089526] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.095541] env[62521]: DEBUG oslo_vmware.api [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319076, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.100270] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319077, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.182195] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.222146] env[62521]: DEBUG oslo_vmware.api [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.549152} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.222443] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.222646] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1044.222832] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1044.223048] env[62521]: INFO nova.compute.manager [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1044.223311] env[62521]: DEBUG oslo.service.loopingcall [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.223517] env[62521]: DEBUG nova.compute.manager [-] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1044.223621] env[62521]: DEBUG nova.network.neutron [-] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1044.360050] env[62521]: DEBUG nova.compute.utils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1044.361580] env[62521]: DEBUG nova.compute.manager [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1044.361785] env[62521]: DEBUG nova.network.neutron [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1044.375846] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319075, 'name': PowerOffVM_Task, 'duration_secs': 0.187085} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.375846] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1044.376623] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d77da5-17ad-48f7-b414-97fbc3217c8d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.397911] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd6fa87-cf1b-4129-bc80-7100c1441142 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.407405] env[62521]: DEBUG nova.policy [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5f378733a3043b3b66f005bc0dce756', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2d7ab38464a4c02a9dc94e04b4793ff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1044.594326] env[62521]: DEBUG oslo_vmware.api [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319076, 'name': ReconfigVM_Task, 'duration_secs': 0.361922} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.598433] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Reconfigured VM instance instance-00000057 to attach disk [datastore2] volume-cfc4d873-e564-4983-a0f5-9f3afecbb1ea/volume-cfc4d873-e564-4983-a0f5-9f3afecbb1ea.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1044.606018] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0898786-4db5-4a52-96c2-5c8ad3bcf961 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.628760] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319077, 'name': ReconfigVM_Task, 'duration_secs': 0.159803} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.630028] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Reconfigured VM instance instance-00000044 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1044.630376] env[62521]: DEBUG oslo_vmware.api [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 1044.630376] env[62521]: value = "task-1319078" [ 1044.630376] env[62521]: _type = "Task" [ 1044.630376] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.631071] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7abe738-cd03-4900-9520-7cf40f6a5234 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.658518] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 7b5afeab-f04e-453a-a3ed-09e3a725e8d6/7b5afeab-f04e-453a-a3ed-09e3a725e8d6.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1044.664868] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0f0c324-5371-4536-b143-3ff1e9aa3fee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.675711] env[62521]: DEBUG oslo_vmware.api [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319078, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.678366] env[62521]: DEBUG nova.network.neutron [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Successfully created port: ee2404e8-7a05-4516-826e-2f5e4adcae45 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1044.685915] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1044.685915] env[62521]: value = "task-1319079" [ 1044.685915] env[62521]: _type = "Task" [ 1044.685915] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.694936] env[62521]: DEBUG oslo_concurrency.lockutils [None req-c4f4174a-59ab-48f6-943b-6a9cd96e0a97 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.264s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.696037] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319079, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.698574] env[62521]: DEBUG nova.compute.manager [req-7b1bbb59-7aa2-4baf-acf3-747a9bf1e052 req-08029676-506f-4aa6-bd29-d13c9b83063b service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Received event network-vif-deleted-9e954d34-4424-459c-a022-fc039668c616 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.698902] env[62521]: INFO nova.compute.manager [req-7b1bbb59-7aa2-4baf-acf3-747a9bf1e052 req-08029676-506f-4aa6-bd29-d13c9b83063b service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Neutron deleted interface 9e954d34-4424-459c-a022-fc039668c616; detaching it from the instance and deleting it from the info cache [ 1044.699125] env[62521]: DEBUG nova.network.neutron [req-7b1bbb59-7aa2-4baf-acf3-747a9bf1e052 req-08029676-506f-4aa6-bd29-d13c9b83063b service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.864857] env[62521]: DEBUG nova.compute.manager [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1044.911019] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1044.911377] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-89b07ddf-9b02-453e-b06b-63eae9065173 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.926134] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1044.926134] env[62521]: value = "task-1319080" [ 1044.926134] env[62521]: _type = "Task" [ 1044.926134] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.939972] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319080, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.053620] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4532dc-8a13-4ce5-9b1d-54a00d9b1345 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.061668] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b838b73b-94e6-4de1-bfad-1e2c435bc0a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.092545] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37ec94a-70f1-4d57-93ac-034fbfbebd8f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.099874] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a168fe9f-178d-4183-bd41-6734e3f796d0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.114339] env[62521]: DEBUG nova.compute.provider_tree [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.142518] env[62521]: DEBUG oslo_vmware.api [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319078, 'name': ReconfigVM_Task, 'duration_secs': 0.137312} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.144419] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282178', 'volume_id': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'name': 'volume-cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c', 'attached_at': '', 'detached_at': '', 'volume_id': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'serial': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1045.181167] env[62521]: DEBUG nova.network.neutron [-] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.198758] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319079, 'name': ReconfigVM_Task, 'duration_secs': 0.261614} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.199881] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 7b5afeab-f04e-453a-a3ed-09e3a725e8d6/7b5afeab-f04e-453a-a3ed-09e3a725e8d6.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1045.200431] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance '7b5afeab-f04e-453a-a3ed-09e3a725e8d6' progress to 50 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1045.208019] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05efc5c2-3dd4-480c-92d8-dbb6e7e017a5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.214711] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949d8ece-27ab-4000-922c-f5592b841d94 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.248802] env[62521]: DEBUG nova.compute.manager [req-7b1bbb59-7aa2-4baf-acf3-747a9bf1e052 req-08029676-506f-4aa6-bd29-d13c9b83063b service nova] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Detach interface failed, port_id=9e954d34-4424-459c-a022-fc039668c616, reason: Instance 11e61e72-a311-4c43-bf53-df82ad9d70f7 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1045.437823] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319080, 'name': CreateSnapshot_Task, 'duration_secs': 0.443795} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.438864] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1045.438975] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68d1b4b-170f-44e3-95be-8b8dbbf041d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.617936] env[62521]: DEBUG nova.scheduler.client.report [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.684513] env[62521]: INFO nova.compute.manager [-] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Took 1.46 seconds to deallocate network for instance. [ 1045.707628] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c54ec8-b3e6-400c-b714-f3c2b21e4645 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.727188] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974dcd6d-eafe-45da-bfac-57a91b7fe3e8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.731453] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.732034] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.732034] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.732305] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.732464] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.749534] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance '7b5afeab-f04e-453a-a3ed-09e3a725e8d6' progress to 67 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1045.752969] env[62521]: INFO nova.compute.manager [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Terminating instance [ 1045.756840] env[62521]: DEBUG nova.compute.manager [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1045.757051] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1045.757816] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a906df7-9b7a-4a0d-8de4-78013d841976 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.765322] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1045.765561] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52edeedd-1cae-4a1e-82bf-eae5cd9b308c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.771394] env[62521]: DEBUG oslo_vmware.api [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1045.771394] env[62521]: value = "task-1319081" [ 1045.771394] env[62521]: _type = "Task" [ 1045.771394] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.779644] env[62521]: DEBUG oslo_vmware.api [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319081, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.878674] env[62521]: DEBUG nova.compute.manager [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1045.906036] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1045.906322] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1045.906483] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.906670] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1045.906822] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.906976] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1045.907257] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1045.907445] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1045.907620] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1045.907788] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1045.908039] env[62521]: DEBUG nova.virt.hardware [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.908919] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea91d58b-7fd5-4baf-aee0-fb162e503f18 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.917059] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425a9741-57ac-40cd-8422-419b75bd3eb4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.960404] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1045.960771] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-34d8a4f7-012f-451f-aec3-e68a1862c965 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.969410] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1045.969410] env[62521]: value = "task-1319082" [ 1045.969410] env[62521]: _type = "Task" [ 1045.969410] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.978350] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319082, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.124171] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.266s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.124798] env[62521]: DEBUG nova.compute.manager [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1046.127796] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.216s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.129838] env[62521]: INFO nova.compute.claims [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1046.184647] env[62521]: DEBUG nova.objects.instance [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'flavor' on Instance uuid 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.191174] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.244215] env[62521]: DEBUG nova.compute.manager [req-2319da97-3bb4-4e11-8a83-92514d7507e0 req-e0321372-3a6f-48ad-9579-ebf29ade8dc8 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Received event network-vif-plugged-ee2404e8-7a05-4516-826e-2f5e4adcae45 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.244215] env[62521]: DEBUG oslo_concurrency.lockutils [req-2319da97-3bb4-4e11-8a83-92514d7507e0 req-e0321372-3a6f-48ad-9579-ebf29ade8dc8 service nova] Acquiring lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.244215] env[62521]: DEBUG oslo_concurrency.lockutils [req-2319da97-3bb4-4e11-8a83-92514d7507e0 req-e0321372-3a6f-48ad-9579-ebf29ade8dc8 service nova] Lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.244215] env[62521]: DEBUG oslo_concurrency.lockutils [req-2319da97-3bb4-4e11-8a83-92514d7507e0 req-e0321372-3a6f-48ad-9579-ebf29ade8dc8 service nova] Lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.244215] env[62521]: DEBUG nova.compute.manager [req-2319da97-3bb4-4e11-8a83-92514d7507e0 req-e0321372-3a6f-48ad-9579-ebf29ade8dc8 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] No waiting events found dispatching network-vif-plugged-ee2404e8-7a05-4516-826e-2f5e4adcae45 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1046.244857] env[62521]: WARNING nova.compute.manager [req-2319da97-3bb4-4e11-8a83-92514d7507e0 req-e0321372-3a6f-48ad-9579-ebf29ade8dc8 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Received unexpected event network-vif-plugged-ee2404e8-7a05-4516-826e-2f5e4adcae45 for instance with vm_state building and task_state spawning. [ 1046.282106] env[62521]: DEBUG oslo_vmware.api [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319081, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.292467] env[62521]: DEBUG nova.network.neutron [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Port 8824a231-01df-41e5-9bea-9071d8b41cb8 binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1046.348443] env[62521]: DEBUG nova.network.neutron [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Successfully updated port: ee2404e8-7a05-4516-826e-2f5e4adcae45 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.480433] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319082, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.636014] env[62521]: DEBUG nova.compute.utils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.638247] env[62521]: DEBUG nova.compute.manager [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1046.638422] env[62521]: DEBUG nova.network.neutron [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1046.679693] env[62521]: DEBUG nova.policy [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5f378733a3043b3b66f005bc0dce756', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2d7ab38464a4c02a9dc94e04b4793ff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1046.689382] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4488ffef-4290-4367-b1ae-620a29d023d4 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.305s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.782218] env[62521]: DEBUG oslo_vmware.api [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319081, 'name': PowerOffVM_Task, 'duration_secs': 0.642881} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.782484] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1046.782659] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1046.782910] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dfeef896-e6eb-4d6c-a451-e26e7a5661a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.847520] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1046.848968] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1046.848968] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleting the datastore file [datastore1] a68d4cb7-bfad-48f6-8aeb-bca04248a0d7 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.848968] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6428d1bd-9480-4875-8bd3-7f3da053b2eb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.854487] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.854487] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.854487] env[62521]: DEBUG nova.network.neutron [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.857463] env[62521]: DEBUG oslo_vmware.api [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1046.857463] env[62521]: value = "task-1319084" [ 1046.857463] env[62521]: _type = "Task" [ 1046.857463] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.866713] env[62521]: DEBUG oslo_vmware.api [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319084, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.952887] env[62521]: DEBUG nova.network.neutron [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Successfully created port: 59e6ef56-abf3-4b61-ab38-86918b4f4ce1 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1046.980293] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319082, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.991853] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.992121] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.139957] env[62521]: DEBUG nova.compute.manager [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1047.308789] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0499c090-7e5e-403b-940f-5b0463b4275c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.313955] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.314210] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.314407] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.327378] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3d6a4e-044f-4a51-9a55-abdba6c34fd4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.367461] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b989d967-cb7a-4461-b403-766493bba8e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.375773] env[62521]: DEBUG oslo_vmware.api [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319084, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218094} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.377814] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1047.378042] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1047.378259] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1047.378535] env[62521]: INFO nova.compute.manager [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1047.378795] env[62521]: DEBUG oslo.service.loopingcall [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.379333] env[62521]: DEBUG nova.compute.manager [-] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1047.379434] env[62521]: DEBUG nova.network.neutron [-] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1047.382197] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96cc861a-f361-451f-aeeb-281f462aa2ab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.396502] env[62521]: DEBUG nova.compute.provider_tree [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.400033] env[62521]: DEBUG nova.network.neutron [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1047.480275] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319082, 'name': CloneVM_Task, 'duration_secs': 1.499205} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.480561] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Created linked-clone VM from snapshot [ 1047.481319] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6b35e5-825d-4500-b2e0-d94a8d0a104b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.488425] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Uploading image 9d0ca44a-c290-4856-a6d8-d11e3d6b1026 {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1047.496656] env[62521]: INFO nova.compute.manager [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Detaching volume e62e324b-ba8a-4465-aa3f-4d13a1020baa [ 1047.517536] env[62521]: DEBUG oslo_vmware.rw_handles [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1047.517536] env[62521]: value = "vm-282180" [ 1047.517536] env[62521]: _type = "VirtualMachine" [ 1047.517536] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1047.517864] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0e49755b-4452-4e4f-b5da-541fe31399b7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.525212] env[62521]: DEBUG oslo_vmware.rw_handles [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lease: (returnval){ [ 1047.525212] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520c9eb0-a293-8b23-15e4-503ffd23e06a" [ 1047.525212] env[62521]: _type = "HttpNfcLease" [ 1047.525212] env[62521]: } obtained for exporting VM: (result){ [ 1047.525212] env[62521]: value = "vm-282180" [ 1047.525212] env[62521]: _type = "VirtualMachine" [ 1047.525212] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1047.525450] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the lease: (returnval){ [ 1047.525450] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520c9eb0-a293-8b23-15e4-503ffd23e06a" [ 1047.525450] env[62521]: _type = "HttpNfcLease" [ 1047.525450] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1047.534313] env[62521]: DEBUG nova.network.neutron [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Updating instance_info_cache with network_info: [{"id": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "address": "fa:16:3e:44:95:f8", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee2404e8-7a", "ovs_interfaceid": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.540081] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1047.540081] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520c9eb0-a293-8b23-15e4-503ffd23e06a" [ 1047.540081] env[62521]: _type = "HttpNfcLease" [ 1047.540081] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1047.540477] env[62521]: DEBUG oslo_vmware.rw_handles [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1047.540477] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520c9eb0-a293-8b23-15e4-503ffd23e06a" [ 1047.540477] env[62521]: _type = "HttpNfcLease" [ 1047.540477] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1047.541207] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ca879c-ba7e-4fcc-8958-4173128ea78c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.545013] env[62521]: INFO nova.virt.block_device [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Attempting to driver detach volume e62e324b-ba8a-4465-aa3f-4d13a1020baa from mountpoint /dev/sdb [ 1047.545013] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1047.545013] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282175', 'volume_id': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'name': 'volume-e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c', 'attached_at': '', 'detached_at': '', 'volume_id': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'serial': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1047.545403] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-814d81b8-c2f3-4ebf-9b93-38ec8a664198 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.553401] env[62521]: DEBUG oslo_vmware.rw_handles [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521ce241-c3ea-9fbe-98e3-05b5cb802e8a/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1047.553573] env[62521]: DEBUG oslo_vmware.rw_handles [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521ce241-c3ea-9fbe-98e3-05b5cb802e8a/disk-0.vmdk for reading. {{(pid=62521) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1047.578656] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c48b8e-5621-4621-a99f-452022434ffc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.640634] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc23104-377f-4c6a-a830-89abeeb6de66 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.669625] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16abcc3a-7f6d-48cb-940c-c926fb3af10b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.672438] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-de0c283d-107c-4ba9-a02d-e0c247a995c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.687165] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] The volume has not been displaced from its original location: [datastore2] volume-e62e324b-ba8a-4465-aa3f-4d13a1020baa/volume-e62e324b-ba8a-4465-aa3f-4d13a1020baa.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1047.692368] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Reconfiguring VM instance instance-00000057 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1047.694793] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47fc12d9-aeb3-4b9f-8033-acda424d649f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.712608] env[62521]: DEBUG oslo_vmware.api [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 1047.712608] env[62521]: value = "task-1319086" [ 1047.712608] env[62521]: _type = "Task" [ 1047.712608] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.723129] env[62521]: DEBUG oslo_vmware.api [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319086, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.899797] env[62521]: DEBUG nova.scheduler.client.report [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.037068] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.037506] env[62521]: DEBUG nova.compute.manager [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Instance network_info: |[{"id": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "address": "fa:16:3e:44:95:f8", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee2404e8-7a", "ovs_interfaceid": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1048.038108] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:95:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e4e52d8a-b086-4333-a5a1-938680a2d2bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee2404e8-7a05-4516-826e-2f5e4adcae45', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1048.045725] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Creating folder: Project (f2d7ab38464a4c02a9dc94e04b4793ff). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1048.046783] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0717e27c-b0ac-4af9-983e-7fe57aa3a328 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.057629] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Created folder: Project (f2d7ab38464a4c02a9dc94e04b4793ff) in parent group-v282025. [ 1048.058503] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Creating folder: Instances. Parent ref: group-v282181. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1048.059185] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-afdfd8ed-55a1-4206-9091-d19faab343ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.068086] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Created folder: Instances in parent group-v282181. [ 1048.068332] env[62521]: DEBUG oslo.service.loopingcall [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.068525] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1048.068723] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3ef4b34-06d5-4d6d-a173-0a0ac0573b15 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.092017] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1048.092017] env[62521]: value = "task-1319089" [ 1048.092017] env[62521]: _type = "Task" [ 1048.092017] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.102534] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319089, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.167858] env[62521]: DEBUG nova.compute.manager [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1048.198573] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.198827] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.198995] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.199222] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.199384] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.199604] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.199884] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.200119] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.200355] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.200579] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.200798] env[62521]: DEBUG nova.virt.hardware [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.201748] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0ae8fb-ed77-4494-8511-fb29ff2500f5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.210231] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6daa0c8d-6e79-4d07-acc2-b32a9e26c68e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.231649] env[62521]: DEBUG oslo_vmware.api [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319086, 'name': ReconfigVM_Task, 'duration_secs': 0.231132} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.232008] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Reconfigured VM instance instance-00000057 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1048.236709] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62e0b7b0-659a-4b79-80e7-36885d5aad8d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.252783] env[62521]: DEBUG oslo_vmware.api [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 1048.252783] env[62521]: value = "task-1319090" [ 1048.252783] env[62521]: _type = "Task" [ 1048.252783] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.261555] env[62521]: DEBUG oslo_vmware.api [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319090, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.276038] env[62521]: DEBUG nova.compute.manager [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Received event network-changed-ee2404e8-7a05-4516-826e-2f5e4adcae45 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.276408] env[62521]: DEBUG nova.compute.manager [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Refreshing instance network info cache due to event network-changed-ee2404e8-7a05-4516-826e-2f5e4adcae45. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1048.276662] env[62521]: DEBUG oslo_concurrency.lockutils [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] Acquiring lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.276874] env[62521]: DEBUG oslo_concurrency.lockutils [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] Acquired lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.277399] env[62521]: DEBUG nova.network.neutron [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Refreshing network info cache for port ee2404e8-7a05-4516-826e-2f5e4adcae45 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1048.395204] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.395513] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.395663] env[62521]: DEBUG nova.network.neutron [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1048.407682] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.408178] env[62521]: DEBUG nova.compute.manager [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1048.412040] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.230s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.414132] env[62521]: INFO nova.compute.claims [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1048.418372] env[62521]: DEBUG nova.network.neutron [-] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.601894] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319089, 'name': CreateVM_Task, 'duration_secs': 0.341991} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.602196] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1048.602987] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.602987] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.603871] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1048.603871] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2e0022d-5ded-43ca-be72-3345c63f048a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.608330] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1048.608330] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520a9634-1e70-70d5-5d60-381a4ebe2ed8" [ 1048.608330] env[62521]: _type = "Task" [ 1048.608330] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.618054] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520a9634-1e70-70d5-5d60-381a4ebe2ed8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.663679] env[62521]: DEBUG nova.network.neutron [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Successfully updated port: 59e6ef56-abf3-4b61-ab38-86918b4f4ce1 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1048.763954] env[62521]: DEBUG oslo_vmware.api [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319090, 'name': ReconfigVM_Task, 'duration_secs': 0.142023} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.766024] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282175', 'volume_id': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'name': 'volume-e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c', 'attached_at': '', 'detached_at': '', 'volume_id': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa', 'serial': 'e62e324b-ba8a-4465-aa3f-4d13a1020baa'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1048.922679] env[62521]: DEBUG nova.compute.utils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1048.925581] env[62521]: INFO nova.compute.manager [-] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Took 1.55 seconds to deallocate network for instance. [ 1048.929019] env[62521]: DEBUG nova.compute.manager [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1048.929019] env[62521]: DEBUG nova.network.neutron [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1048.977695] env[62521]: DEBUG nova.policy [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76d85009c2c1416b9ea049003e6d96a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c533bd4917b4466e9c5c76a9caf8e807', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1049.120321] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520a9634-1e70-70d5-5d60-381a4ebe2ed8, 'name': SearchDatastore_Task, 'duration_secs': 0.017866} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.122844] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.123134] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1049.123422] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.123585] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.123810] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1049.126719] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92857c77-04b1-413e-84ce-ad5a817be310 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.135762] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1049.136179] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1049.136958] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cc002cd-545c-44df-bc6a-0daafa30ac30 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.151116] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1049.151116] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520d03bf-0ed5-8553-6317-1f7cb41a953c" [ 1049.151116] env[62521]: _type = "Task" [ 1049.151116] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.159942] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520d03bf-0ed5-8553-6317-1f7cb41a953c, 'name': SearchDatastore_Task, 'duration_secs': 0.011386} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.160800] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db92f176-f31d-458b-9d02-349992e199ea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.167186] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "refresh_cache-4021ce0d-69ee-4c57-bb18-fd77f339ed72" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.167313] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "refresh_cache-4021ce0d-69ee-4c57-bb18-fd77f339ed72" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.167419] env[62521]: DEBUG nova.network.neutron [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1049.174124] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1049.174124] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e31b79-b59d-d305-3411-8e994b617995" [ 1049.174124] env[62521]: _type = "Task" [ 1049.174124] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.174441] env[62521]: DEBUG nova.network.neutron [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Updated VIF entry in instance network info cache for port ee2404e8-7a05-4516-826e-2f5e4adcae45. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1049.174792] env[62521]: DEBUG nova.network.neutron [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Updating instance_info_cache with network_info: [{"id": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "address": "fa:16:3e:44:95:f8", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee2404e8-7a", "ovs_interfaceid": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.182111] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e31b79-b59d-d305-3411-8e994b617995, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.275484] env[62521]: DEBUG nova.network.neutron [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Successfully created port: e02e66bc-8bdd-4eed-81a7-612a3d86e2f7 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1049.283826] env[62521]: DEBUG nova.network.neutron [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [{"id": "8824a231-01df-41e5-9bea-9071d8b41cb8", "address": "fa:16:3e:2e:cc:1b", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8824a231-01", "ovs_interfaceid": "8824a231-01df-41e5-9bea-9071d8b41cb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.313595] env[62521]: DEBUG nova.objects.instance [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'flavor' on Instance uuid 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.429642] env[62521]: DEBUG nova.compute.manager [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1049.438324] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.598926] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db074e59-0c16-4194-aac0-b66302ad061e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.608339] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d3abd9-44e4-46e8-99af-60a2361031a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.638448] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99e37512-8dc5-4f6b-8e84-e037f6328167 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.646878] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f26b380-88f3-4f95-aa3c-d3c45e9da6a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.660896] env[62521]: DEBUG nova.compute.provider_tree [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.677268] env[62521]: DEBUG oslo_concurrency.lockutils [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] Releasing lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.677545] env[62521]: DEBUG nova.compute.manager [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Received event network-vif-deleted-add56f48-9368-4591-b8b6-cc7be1121d28 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.677730] env[62521]: INFO nova.compute.manager [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Neutron deleted interface add56f48-9368-4591-b8b6-cc7be1121d28; detaching it from the instance and deleting it from the info cache [ 1049.677908] env[62521]: DEBUG nova.network.neutron [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.684726] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e31b79-b59d-d305-3411-8e994b617995, 'name': SearchDatastore_Task, 'duration_secs': 0.013962} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.684979] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.685255] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 703ff423-98b1-4942-b6fd-8b95fe57bd0a/703ff423-98b1-4942-b6fd-8b95fe57bd0a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1049.685502] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ffa0b3e-ba79-462c-92df-d58af8774ccc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.694142] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1049.694142] env[62521]: value = "task-1319091" [ 1049.694142] env[62521]: _type = "Task" [ 1049.694142] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.702942] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319091, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.704497] env[62521]: DEBUG nova.network.neutron [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1049.789146] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.857310] env[62521]: DEBUG nova.network.neutron [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Updating instance_info_cache with network_info: [{"id": "59e6ef56-abf3-4b61-ab38-86918b4f4ce1", "address": "fa:16:3e:59:05:3e", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59e6ef56-ab", "ovs_interfaceid": "59e6ef56-abf3-4b61-ab38-86918b4f4ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.165100] env[62521]: DEBUG nova.scheduler.client.report [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.183975] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-69a67e18-4b84-487e-9ef2-b0ce5e8105b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.198117] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076ba60d-b837-4129-ace9-ed23b73783dd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.224383] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319091, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.241807] env[62521]: DEBUG nova.compute.manager [req-7dda30fb-ce63-48be-b279-d745884b08e2 req-8fb31b77-c257-48a4-8af5-4697a31057b2 service nova] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Detach interface failed, port_id=add56f48-9368-4591-b8b6-cc7be1121d28, reason: Instance a68d4cb7-bfad-48f6-8aeb-bca04248a0d7 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1050.263789] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.315127] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e823a498-04ff-4858-9c3c-225b80609bbb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.319896] env[62521]: DEBUG nova.compute.manager [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Received event network-vif-plugged-59e6ef56-abf3-4b61-ab38-86918b4f4ce1 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1050.320141] env[62521]: DEBUG oslo_concurrency.lockutils [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] Acquiring lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.320354] env[62521]: DEBUG oslo_concurrency.lockutils [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] Lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.320530] env[62521]: DEBUG oslo_concurrency.lockutils [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] Lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.320705] env[62521]: DEBUG nova.compute.manager [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] No waiting events found dispatching network-vif-plugged-59e6ef56-abf3-4b61-ab38-86918b4f4ce1 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1050.320876] env[62521]: WARNING nova.compute.manager [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Received unexpected event network-vif-plugged-59e6ef56-abf3-4b61-ab38-86918b4f4ce1 for instance with vm_state building and task_state spawning. [ 1050.321053] env[62521]: DEBUG nova.compute.manager [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Received event network-changed-59e6ef56-abf3-4b61-ab38-86918b4f4ce1 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1050.321220] env[62521]: DEBUG nova.compute.manager [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Refreshing instance network info cache due to event network-changed-59e6ef56-abf3-4b61-ab38-86918b4f4ce1. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1050.321389] env[62521]: DEBUG oslo_concurrency.lockutils [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] Acquiring lock "refresh_cache-4021ce0d-69ee-4c57-bb18-fd77f339ed72" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.322414] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ea016d6a-1ae6-4875-b106-f31093c9285b tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.330s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.323843] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.060s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.341534] env[62521]: INFO nova.compute.manager [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Detaching volume cfc4d873-e564-4983-a0f5-9f3afecbb1ea [ 1050.343849] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b47451-fee4-4766-b6c3-3a578d10fe41 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.355494] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance '7b5afeab-f04e-453a-a3ed-09e3a725e8d6' progress to 83 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1050.359830] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "refresh_cache-4021ce0d-69ee-4c57-bb18-fd77f339ed72" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.359830] env[62521]: DEBUG nova.compute.manager [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Instance network_info: |[{"id": "59e6ef56-abf3-4b61-ab38-86918b4f4ce1", "address": "fa:16:3e:59:05:3e", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59e6ef56-ab", "ovs_interfaceid": "59e6ef56-abf3-4b61-ab38-86918b4f4ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1050.360631] env[62521]: DEBUG oslo_concurrency.lockutils [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] Acquired lock "refresh_cache-4021ce0d-69ee-4c57-bb18-fd77f339ed72" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.360631] env[62521]: DEBUG nova.network.neutron [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Refreshing network info cache for port 59e6ef56-abf3-4b61-ab38-86918b4f4ce1 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1050.361524] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:05:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e4e52d8a-b086-4333-a5a1-938680a2d2bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '59e6ef56-abf3-4b61-ab38-86918b4f4ce1', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1050.368836] env[62521]: DEBUG oslo.service.loopingcall [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.369734] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1050.369965] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03e93ccc-9850-419e-a466-aa171f33e3da {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.385533] env[62521]: INFO nova.virt.block_device [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Attempting to driver detach volume cfc4d873-e564-4983-a0f5-9f3afecbb1ea from mountpoint /dev/sdc [ 1050.385783] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1050.385985] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282178', 'volume_id': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'name': 'volume-cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c', 'attached_at': '', 'detached_at': '', 'volume_id': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'serial': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1050.387383] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b7927e-3bfc-41bf-ba85-94c70d4f3583 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.412091] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca6c06f-a338-4cde-ba47-4fed459746bf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.415187] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1050.415187] env[62521]: value = "task-1319092" [ 1050.415187] env[62521]: _type = "Task" [ 1050.415187] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.421514] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4d6659-26be-4ae3-9467-52eca3f29bed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.429730] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319092, 'name': CreateVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.448048] env[62521]: DEBUG nova.compute.manager [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1050.450931] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e950e6-ac7b-4f0e-bb42-ec38d5691366 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.467464] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] The volume has not been displaced from its original location: [datastore2] volume-cfc4d873-e564-4983-a0f5-9f3afecbb1ea/volume-cfc4d873-e564-4983-a0f5-9f3afecbb1ea.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1050.472873] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Reconfiguring VM instance instance-00000057 to detach disk 2002 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1050.475377] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc588589-0bdc-44f6-90dc-fe96e47e7b12 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.495291] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1050.495564] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1050.495727] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1050.495926] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1050.496129] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1050.496299] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1050.496515] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1050.496681] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1050.496857] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1050.497040] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1050.497228] env[62521]: DEBUG nova.virt.hardware [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1050.498132] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9fd15a-7b98-4cf0-befc-ac7df7c6a50e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.504691] env[62521]: DEBUG oslo_vmware.api [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 1050.504691] env[62521]: value = "task-1319093" [ 1050.504691] env[62521]: _type = "Task" [ 1050.504691] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.511416] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8efe05-5e7a-4376-9d3b-8c3fd1c8abb1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.518343] env[62521]: DEBUG oslo_vmware.api [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319093, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.671346] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.671852] env[62521]: DEBUG nova.compute.manager [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1050.674838] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.484s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.675064] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.677298] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.239s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.677530] env[62521]: DEBUG nova.objects.instance [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lazy-loading 'resources' on Instance uuid a68d4cb7-bfad-48f6-8aeb-bca04248a0d7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.696170] env[62521]: INFO nova.scheduler.client.report [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleted allocations for instance 11e61e72-a311-4c43-bf53-df82ad9d70f7 [ 1050.703340] env[62521]: DEBUG nova.compute.manager [req-7394ea8c-f098-42eb-ae3c-e900cc168802 req-82e98447-9850-49de-b167-25ecd649bc61 service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Received event network-vif-plugged-e02e66bc-8bdd-4eed-81a7-612a3d86e2f7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1050.703611] env[62521]: DEBUG oslo_concurrency.lockutils [req-7394ea8c-f098-42eb-ae3c-e900cc168802 req-82e98447-9850-49de-b167-25ecd649bc61 service nova] Acquiring lock "bc7d76b1-76b2-49a7-8188-0672312767a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.703854] env[62521]: DEBUG oslo_concurrency.lockutils [req-7394ea8c-f098-42eb-ae3c-e900cc168802 req-82e98447-9850-49de-b167-25ecd649bc61 service nova] Lock "bc7d76b1-76b2-49a7-8188-0672312767a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.704061] env[62521]: DEBUG oslo_concurrency.lockutils [req-7394ea8c-f098-42eb-ae3c-e900cc168802 req-82e98447-9850-49de-b167-25ecd649bc61 service nova] Lock "bc7d76b1-76b2-49a7-8188-0672312767a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.704252] env[62521]: DEBUG nova.compute.manager [req-7394ea8c-f098-42eb-ae3c-e900cc168802 req-82e98447-9850-49de-b167-25ecd649bc61 service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] No waiting events found dispatching network-vif-plugged-e02e66bc-8bdd-4eed-81a7-612a3d86e2f7 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1050.704469] env[62521]: WARNING nova.compute.manager [req-7394ea8c-f098-42eb-ae3c-e900cc168802 req-82e98447-9850-49de-b167-25ecd649bc61 service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Received unexpected event network-vif-plugged-e02e66bc-8bdd-4eed-81a7-612a3d86e2f7 for instance with vm_state building and task_state spawning. [ 1050.727322] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319091, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.604729} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.727539] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 703ff423-98b1-4942-b6fd-8b95fe57bd0a/703ff423-98b1-4942-b6fd-8b95fe57bd0a.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1050.727762] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1050.728468] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-28b39049-384a-423d-8bc3-66d82d2e421d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.738354] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1050.738354] env[62521]: value = "task-1319094" [ 1050.738354] env[62521]: _type = "Task" [ 1050.738354] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.749723] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319094, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.862840] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.863278] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9abb650-4e6d-4b31-9995-5199be5573d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.873507] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1050.873507] env[62521]: value = "task-1319095" [ 1050.873507] env[62521]: _type = "Task" [ 1050.873507] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.883837] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319095, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.925692] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319092, 'name': CreateVM_Task, 'duration_secs': 0.474194} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.925916] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1050.926685] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.926866] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.927266] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1050.927581] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d40b5ab3-3dfd-45f8-97d4-dd325542f6f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.934449] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1050.934449] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b54e01-4e76-24e1-3808-4b3ec7caff31" [ 1050.934449] env[62521]: _type = "Task" [ 1050.934449] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.944122] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b54e01-4e76-24e1-3808-4b3ec7caff31, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.018058] env[62521]: DEBUG oslo_vmware.api [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319093, 'name': ReconfigVM_Task, 'duration_secs': 0.310716} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.018475] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Reconfigured VM instance instance-00000057 to detach disk 2002 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1051.029882] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40c19947-cfb6-46dd-b60b-e30e586eb8f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.047601] env[62521]: DEBUG oslo_vmware.api [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 1051.047601] env[62521]: value = "task-1319096" [ 1051.047601] env[62521]: _type = "Task" [ 1051.047601] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.058489] env[62521]: DEBUG oslo_vmware.api [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319096, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.134561] env[62521]: DEBUG nova.network.neutron [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Updated VIF entry in instance network info cache for port 59e6ef56-abf3-4b61-ab38-86918b4f4ce1. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1051.134955] env[62521]: DEBUG nova.network.neutron [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Updating instance_info_cache with network_info: [{"id": "59e6ef56-abf3-4b61-ab38-86918b4f4ce1", "address": "fa:16:3e:59:05:3e", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59e6ef56-ab", "ovs_interfaceid": "59e6ef56-abf3-4b61-ab38-86918b4f4ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.181596] env[62521]: DEBUG nova.compute.utils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1051.185901] env[62521]: DEBUG nova.compute.manager [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1051.186168] env[62521]: DEBUG nova.network.neutron [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1051.207938] env[62521]: DEBUG oslo_concurrency.lockutils [None req-9e084682-e15e-43d6-8b15-a3998bbcb6d7 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "11e61e72-a311-4c43-bf53-df82ad9d70f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.630s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.259759] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319094, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077904} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.261729] env[62521]: DEBUG nova.policy [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ce6cad0be5f455eab09b1d348268329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '961924204a584b61a4a85a761821910d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1051.263038] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1051.264062] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e86edc8-d52d-4803-addf-dfbfcd3e68aa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.296618] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 703ff423-98b1-4942-b6fd-8b95fe57bd0a/703ff423-98b1-4942-b6fd-8b95fe57bd0a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1051.297636] env[62521]: DEBUG nova.network.neutron [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Successfully updated port: e02e66bc-8bdd-4eed-81a7-612a3d86e2f7 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1051.301309] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7ddc62b-5ebb-4d4c-a36a-1f04584226d0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.323626] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1051.323626] env[62521]: value = "task-1319097" [ 1051.323626] env[62521]: _type = "Task" [ 1051.323626] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.335533] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319097, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.384398] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319095, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.421076] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e5e42c-8ff9-47e5-acc0-9a8c9c9eeb21 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.430055] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c0a030c-10aa-4095-b7d9-0397cb802b6a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.469889] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b7f2d0-2e3f-4420-83d1-40bd9608f2e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.474247] env[62521]: DEBUG nova.compute.manager [req-ec2cf98f-558b-4749-956e-184532ae97eb req-a37fccac-f48d-4991-bf56-ef248c61edc0 service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Received event network-changed-e02e66bc-8bdd-4eed-81a7-612a3d86e2f7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.474473] env[62521]: DEBUG nova.compute.manager [req-ec2cf98f-558b-4749-956e-184532ae97eb req-a37fccac-f48d-4991-bf56-ef248c61edc0 service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Refreshing instance network info cache due to event network-changed-e02e66bc-8bdd-4eed-81a7-612a3d86e2f7. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1051.474695] env[62521]: DEBUG oslo_concurrency.lockutils [req-ec2cf98f-558b-4749-956e-184532ae97eb req-a37fccac-f48d-4991-bf56-ef248c61edc0 service nova] Acquiring lock "refresh_cache-bc7d76b1-76b2-49a7-8188-0672312767a1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.474882] env[62521]: DEBUG oslo_concurrency.lockutils [req-ec2cf98f-558b-4749-956e-184532ae97eb req-a37fccac-f48d-4991-bf56-ef248c61edc0 service nova] Acquired lock "refresh_cache-bc7d76b1-76b2-49a7-8188-0672312767a1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.475097] env[62521]: DEBUG nova.network.neutron [req-ec2cf98f-558b-4749-956e-184532ae97eb req-a37fccac-f48d-4991-bf56-ef248c61edc0 service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Refreshing network info cache for port e02e66bc-8bdd-4eed-81a7-612a3d86e2f7 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1051.480886] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b54e01-4e76-24e1-3808-4b3ec7caff31, 'name': SearchDatastore_Task, 'duration_secs': 0.025033} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.481588] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.481886] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.482194] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.482505] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.482727] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.483356] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f197429d-2446-4003-a36c-a3868c32781a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.499728] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3305caeb-6385-49b4-857e-91c04207a9f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.505050] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.505271] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1051.506407] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43ac468c-121e-4296-82dd-e498fb06b542 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.518075] env[62521]: DEBUG nova.compute.provider_tree [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.526020] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1051.526020] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526e3c8c-628f-49d6-cfa1-5f01239f3d08" [ 1051.526020] env[62521]: _type = "Task" [ 1051.526020] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.532414] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526e3c8c-628f-49d6-cfa1-5f01239f3d08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.558765] env[62521]: DEBUG oslo_vmware.api [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319096, 'name': ReconfigVM_Task, 'duration_secs': 0.175081} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.559296] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282178', 'volume_id': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'name': 'volume-cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c', 'attached_at': '', 'detached_at': '', 'volume_id': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea', 'serial': 'cfc4d873-e564-4983-a0f5-9f3afecbb1ea'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1051.637702] env[62521]: DEBUG oslo_concurrency.lockutils [req-384fd1c6-af0f-43b9-8592-8e1e0422eb16 req-910a21cc-a767-45f3-830f-2ccd44007b17 service nova] Releasing lock "refresh_cache-4021ce0d-69ee-4c57-bb18-fd77f339ed72" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.689285] env[62521]: DEBUG nova.compute.manager [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1051.791166] env[62521]: DEBUG nova.network.neutron [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Successfully created port: 2beae869-6010-48ba-8e25-6b392dd2ec1f {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1051.816903] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "refresh_cache-bc7d76b1-76b2-49a7-8188-0672312767a1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.833413] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319097, 'name': ReconfigVM_Task, 'duration_secs': 0.293311} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.833694] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 703ff423-98b1-4942-b6fd-8b95fe57bd0a/703ff423-98b1-4942-b6fd-8b95fe57bd0a.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.834334] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5509bd1c-fa0e-4298-ab94-fa2c33a47b95 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.841879] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1051.841879] env[62521]: value = "task-1319098" [ 1051.841879] env[62521]: _type = "Task" [ 1051.841879] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.850079] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319098, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.884163] env[62521]: DEBUG oslo_vmware.api [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319095, 'name': PowerOnVM_Task, 'duration_secs': 0.64674} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.884475] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1051.884682] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce3ab6b-55e9-43fb-9087-432565b0ca95 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance '7b5afeab-f04e-453a-a3ed-09e3a725e8d6' progress to 100 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1052.015373] env[62521]: DEBUG nova.network.neutron [req-ec2cf98f-558b-4749-956e-184532ae97eb req-a37fccac-f48d-4991-bf56-ef248c61edc0 service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1052.022018] env[62521]: DEBUG nova.scheduler.client.report [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.035505] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526e3c8c-628f-49d6-cfa1-5f01239f3d08, 'name': SearchDatastore_Task, 'duration_secs': 0.023994} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.036229] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80bb45b0-8e6a-4d20-be64-b8e8207dbeb9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.042710] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1052.042710] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d475c9-7625-627e-fc87-4ebee629f3e1" [ 1052.042710] env[62521]: _type = "Task" [ 1052.042710] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.050848] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d475c9-7625-627e-fc87-4ebee629f3e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.094439] env[62521]: DEBUG nova.network.neutron [req-ec2cf98f-558b-4749-956e-184532ae97eb req-a37fccac-f48d-4991-bf56-ef248c61edc0 service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.100952] env[62521]: DEBUG nova.objects.instance [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'flavor' on Instance uuid 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.352529] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319098, 'name': Rename_Task, 'duration_secs': 0.139474} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.352815] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1052.353137] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-224b217d-97bc-49a4-98f9-a72a7c41ceb5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.362571] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1052.362571] env[62521]: value = "task-1319099" [ 1052.362571] env[62521]: _type = "Task" [ 1052.362571] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.369900] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319099, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.373924] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.374180] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.529593] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.852s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.554079] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d475c9-7625-627e-fc87-4ebee629f3e1, 'name': SearchDatastore_Task, 'duration_secs': 0.071744} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.555130] env[62521]: INFO nova.scheduler.client.report [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleted allocations for instance a68d4cb7-bfad-48f6-8aeb-bca04248a0d7 [ 1052.556227] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.556504] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 4021ce0d-69ee-4c57-bb18-fd77f339ed72/4021ce0d-69ee-4c57-bb18-fd77f339ed72.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1052.559144] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6fe6b51-76eb-485b-9456-dbc4c2a9de0d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.568459] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1052.568459] env[62521]: value = "task-1319100" [ 1052.568459] env[62521]: _type = "Task" [ 1052.568459] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.577187] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319100, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.597029] env[62521]: DEBUG oslo_concurrency.lockutils [req-ec2cf98f-558b-4749-956e-184532ae97eb req-a37fccac-f48d-4991-bf56-ef248c61edc0 service nova] Releasing lock "refresh_cache-bc7d76b1-76b2-49a7-8188-0672312767a1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.597379] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "refresh_cache-bc7d76b1-76b2-49a7-8188-0672312767a1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.597543] env[62521]: DEBUG nova.network.neutron [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1052.699077] env[62521]: DEBUG nova.compute.manager [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1052.724814] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1052.725110] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1052.725290] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1052.725490] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1052.725650] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1052.725818] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1052.726048] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1052.726237] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1052.726425] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1052.726602] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1052.726789] env[62521]: DEBUG nova.virt.hardware [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1052.727712] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701a593b-365f-406c-bc41-10a1446fe18e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.736707] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa54a80-d88d-46ca-ad57-c1ba4b170191 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.872733] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319099, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.876358] env[62521]: DEBUG nova.compute.manager [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1053.064947] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8fad833c-d597-4c1e-8789-3e52d56eb260 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "a68d4cb7-bfad-48f6-8aeb-bca04248a0d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.333s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.079877] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319100, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.107911] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2932dfb6-0c05-430d-98ab-407b65a2d7c9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 2.784s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.134240] env[62521]: DEBUG nova.network.neutron [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1053.342068] env[62521]: DEBUG nova.network.neutron [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Updating instance_info_cache with network_info: [{"id": "e02e66bc-8bdd-4eed-81a7-612a3d86e2f7", "address": "fa:16:3e:66:3b:60", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02e66bc-8b", "ovs_interfaceid": "e02e66bc-8bdd-4eed-81a7-612a3d86e2f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.352763] env[62521]: DEBUG nova.compute.manager [req-f4efe278-3ac6-43f6-9a16-1b48817b9ea5 req-f74fe853-3c9e-4a42-86cb-3d271e0c203c service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received event network-vif-plugged-2beae869-6010-48ba-8e25-6b392dd2ec1f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.353042] env[62521]: DEBUG oslo_concurrency.lockutils [req-f4efe278-3ac6-43f6-9a16-1b48817b9ea5 req-f74fe853-3c9e-4a42-86cb-3d271e0c203c service nova] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.353276] env[62521]: DEBUG oslo_concurrency.lockutils [req-f4efe278-3ac6-43f6-9a16-1b48817b9ea5 req-f74fe853-3c9e-4a42-86cb-3d271e0c203c service nova] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.353455] env[62521]: DEBUG oslo_concurrency.lockutils [req-f4efe278-3ac6-43f6-9a16-1b48817b9ea5 req-f74fe853-3c9e-4a42-86cb-3d271e0c203c service nova] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.353628] env[62521]: DEBUG nova.compute.manager [req-f4efe278-3ac6-43f6-9a16-1b48817b9ea5 req-f74fe853-3c9e-4a42-86cb-3d271e0c203c service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] No waiting events found dispatching network-vif-plugged-2beae869-6010-48ba-8e25-6b392dd2ec1f {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1053.353798] env[62521]: WARNING nova.compute.manager [req-f4efe278-3ac6-43f6-9a16-1b48817b9ea5 req-f74fe853-3c9e-4a42-86cb-3d271e0c203c service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received unexpected event network-vif-plugged-2beae869-6010-48ba-8e25-6b392dd2ec1f for instance with vm_state building and task_state spawning. [ 1053.372803] env[62521]: DEBUG oslo_vmware.api [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319099, 'name': PowerOnVM_Task, 'duration_secs': 0.697048} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.373076] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1053.373288] env[62521]: INFO nova.compute.manager [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Took 7.49 seconds to spawn the instance on the hypervisor. [ 1053.373547] env[62521]: DEBUG nova.compute.manager [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.374338] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e855eafc-15da-4da0-b937-cca46428296b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.405469] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.408221] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.408221] env[62521]: INFO nova.compute.claims [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1053.523929] env[62521]: DEBUG nova.network.neutron [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Successfully updated port: 2beae869-6010-48ba-8e25-6b392dd2ec1f {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1053.578971] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319100, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.686989] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.686989] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.686989] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.686989] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.686989] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.692952] env[62521]: INFO nova.compute.manager [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Terminating instance [ 1053.701017] env[62521]: DEBUG nova.compute.manager [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1053.701017] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1053.701017] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c25451-d7dd-4b69-a044-161b61f533d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.708810] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1053.709229] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7c72460-f487-40df-b754-88135e09308f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.716568] env[62521]: DEBUG oslo_vmware.api [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 1053.716568] env[62521]: value = "task-1319101" [ 1053.716568] env[62521]: _type = "Task" [ 1053.716568] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.724361] env[62521]: DEBUG oslo_vmware.api [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319101, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.847024] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "refresh_cache-bc7d76b1-76b2-49a7-8188-0672312767a1" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.847024] env[62521]: DEBUG nova.compute.manager [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Instance network_info: |[{"id": "e02e66bc-8bdd-4eed-81a7-612a3d86e2f7", "address": "fa:16:3e:66:3b:60", "network": {"id": "bde63270-20b8-4f2a-9fab-9d05f7b64da7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1925094567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c533bd4917b4466e9c5c76a9caf8e807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "84aee122-f630-43c5-9cc1-3a38d3819c82", "external-id": "nsx-vlan-transportzone-816", "segmentation_id": 816, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02e66bc-8b", "ovs_interfaceid": "e02e66bc-8bdd-4eed-81a7-612a3d86e2f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1053.847024] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:3b:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '84aee122-f630-43c5-9cc1-3a38d3819c82', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e02e66bc-8bdd-4eed-81a7-612a3d86e2f7', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1053.860368] env[62521]: DEBUG oslo.service.loopingcall [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.860918] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1053.861361] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b08766a-ac17-412c-a95f-67bfe3f2a3cf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.885463] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1053.885463] env[62521]: value = "task-1319102" [ 1053.885463] env[62521]: _type = "Task" [ 1053.885463] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.908326] env[62521]: INFO nova.compute.manager [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Took 13.80 seconds to build instance. [ 1053.909585] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319102, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.028757] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.028757] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.028757] env[62521]: DEBUG nova.network.neutron [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1054.079732] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319100, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.227142] env[62521]: DEBUG oslo_vmware.api [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319101, 'name': PowerOffVM_Task, 'duration_secs': 0.199401} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.227436] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1054.227682] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1054.227949] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6aa3e737-986c-484c-acc9-d6a90a28ce86 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.316914] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1054.317233] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1054.317473] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Deleting the datastore file [datastore1] 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.317814] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c9293fb3-096f-4366-bc47-41bba8ead07d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.326496] env[62521]: DEBUG oslo_vmware.api [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for the task: (returnval){ [ 1054.326496] env[62521]: value = "task-1319104" [ 1054.326496] env[62521]: _type = "Task" [ 1054.326496] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.335462] env[62521]: DEBUG oslo_vmware.api [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319104, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.395969] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319102, 'name': CreateVM_Task, 'duration_secs': 0.440249} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.396250] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1054.396992] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.397388] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.397729] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1054.397995] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5181ba4-a15d-4fbd-9021-6eedd4fd09dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.402806] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1054.402806] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52384914-dbe5-4017-f4c2-ae00935b7b28" [ 1054.402806] env[62521]: _type = "Task" [ 1054.402806] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.411516] env[62521]: DEBUG oslo_concurrency.lockutils [None req-681f524e-1309-4b19-8a34-7b8131f42bb7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.308s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.411945] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52384914-dbe5-4017-f4c2-ae00935b7b28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.582359] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319100, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.631101} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.582600] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 4021ce0d-69ee-4c57-bb18-fd77f339ed72/4021ce0d-69ee-4c57-bb18-fd77f339ed72.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1054.582818] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1054.583096] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e112ae9b-afd4-4c9f-9b3c-93079405f308 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.590038] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1054.590038] env[62521]: value = "task-1319105" [ 1054.590038] env[62521]: _type = "Task" [ 1054.590038] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.599513] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319105, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.600345] env[62521]: DEBUG nova.network.neutron [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1054.606714] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df24583e-e12b-4dd6-b511-3d5ded353473 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.615395] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f0a759-939c-4201-a488-c959a84da3d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.651048] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89b53ce-75ea-4c68-acbd-ce933c94f6c3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.658270] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811aa45c-eaad-4ebd-bdb6-3a61de4359e3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.671797] env[62521]: DEBUG nova.compute.provider_tree [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.750124] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.750540] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.750605] env[62521]: DEBUG nova.compute.manager [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Going to confirm migration 6 {{(pid=62521) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1054.837833] env[62521]: DEBUG oslo_vmware.api [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Task: {'id': task-1319104, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.492591} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.838489] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1054.838684] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1054.838870] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1054.839068] env[62521]: INFO nova.compute.manager [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1054.839316] env[62521]: DEBUG oslo.service.loopingcall [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1054.839517] env[62521]: DEBUG nova.compute.manager [-] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1054.839613] env[62521]: DEBUG nova.network.neutron [-] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1054.848225] env[62521]: DEBUG nova.network.neutron [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.912784] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52384914-dbe5-4017-f4c2-ae00935b7b28, 'name': SearchDatastore_Task, 'duration_secs': 0.029556} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.912923] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.914173] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1054.914173] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.914173] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.914173] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1054.914173] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fdcd302a-1c0b-410e-85de-8addffa49c84 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.922055] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1054.922254] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1054.922990] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65c77298-c671-41d1-88b0-9008e79ed490 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.929387] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1054.929387] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52679c93-9032-1198-754f-1619e4aecb52" [ 1054.929387] env[62521]: _type = "Task" [ 1054.929387] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.937297] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52679c93-9032-1198-754f-1619e4aecb52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.100013] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319105, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.173557} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.100192] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1055.101166] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3eadff-48a2-455e-88ab-b9830a79d5ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.129760] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 4021ce0d-69ee-4c57-bb18-fd77f339ed72/4021ce0d-69ee-4c57-bb18-fd77f339ed72.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1055.130438] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5d02ea9-2c91-4cff-a4a5-4147785b250d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.151286] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1055.151286] env[62521]: value = "task-1319106" [ 1055.151286] env[62521]: _type = "Task" [ 1055.151286] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.162442] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319106, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.176813] env[62521]: DEBUG nova.scheduler.client.report [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.295106] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.295222] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.295329] env[62521]: DEBUG nova.network.neutron [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1055.295524] env[62521]: DEBUG nova.objects.instance [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'info_cache' on Instance uuid 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.351065] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.352034] env[62521]: DEBUG nova.compute.manager [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Instance network_info: |[{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1055.352034] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:4e:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92cdccfd-4b10-4024-b724-5f22792dd4de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2beae869-6010-48ba-8e25-6b392dd2ec1f', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1055.361662] env[62521]: DEBUG oslo.service.loopingcall [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.362434] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1055.362714] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2a225243-0315-4c46-aa0a-2c196262c6f1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.381203] env[62521]: DEBUG nova.compute.manager [req-c6cb6c0c-8f7f-43d1-904f-4d79322fb17a req-a37b07a4-dc87-436b-b122-5ae283a6b849 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Received event network-vif-deleted-dee03f5c-2442-438b-8786-6f3e0c35f8ab {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.381484] env[62521]: INFO nova.compute.manager [req-c6cb6c0c-8f7f-43d1-904f-4d79322fb17a req-a37b07a4-dc87-436b-b122-5ae283a6b849 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Neutron deleted interface dee03f5c-2442-438b-8786-6f3e0c35f8ab; detaching it from the instance and deleting it from the info cache [ 1055.381773] env[62521]: DEBUG nova.network.neutron [req-c6cb6c0c-8f7f-43d1-904f-4d79322fb17a req-a37b07a4-dc87-436b-b122-5ae283a6b849 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.388147] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1055.388147] env[62521]: value = "task-1319107" [ 1055.388147] env[62521]: _type = "Task" [ 1055.388147] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.400297] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319107, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.414464] env[62521]: DEBUG nova.compute.manager [req-a942089e-fbc0-41fd-9b0b-25f6dd062bee req-fcfd2222-2b04-4246-84ac-3c9d7c25cf93 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received event network-changed-2beae869-6010-48ba-8e25-6b392dd2ec1f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.414740] env[62521]: DEBUG nova.compute.manager [req-a942089e-fbc0-41fd-9b0b-25f6dd062bee req-fcfd2222-2b04-4246-84ac-3c9d7c25cf93 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Refreshing instance network info cache due to event network-changed-2beae869-6010-48ba-8e25-6b392dd2ec1f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1055.415697] env[62521]: DEBUG oslo_concurrency.lockutils [req-a942089e-fbc0-41fd-9b0b-25f6dd062bee req-fcfd2222-2b04-4246-84ac-3c9d7c25cf93 service nova] Acquiring lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.415697] env[62521]: DEBUG oslo_concurrency.lockutils [req-a942089e-fbc0-41fd-9b0b-25f6dd062bee req-fcfd2222-2b04-4246-84ac-3c9d7c25cf93 service nova] Acquired lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.415697] env[62521]: DEBUG nova.network.neutron [req-a942089e-fbc0-41fd-9b0b-25f6dd062bee req-fcfd2222-2b04-4246-84ac-3c9d7c25cf93 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Refreshing network info cache for port 2beae869-6010-48ba-8e25-6b392dd2ec1f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1055.441218] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52679c93-9032-1198-754f-1619e4aecb52, 'name': SearchDatastore_Task, 'duration_secs': 0.022495} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.442573] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62035fbd-657d-4b44-b4eb-7a6351355233 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.448397] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1055.448397] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5218f426-58f4-e753-5447-7e1e2c1f8e7c" [ 1055.448397] env[62521]: _type = "Task" [ 1055.448397] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.458820] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5218f426-58f4-e753-5447-7e1e2c1f8e7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.662338] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319106, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.681548] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.682168] env[62521]: DEBUG nova.compute.manager [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1055.829320] env[62521]: DEBUG nova.network.neutron [-] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.885205] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08f7449f-9807-4102-8f06-e639c694775a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.896538] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-972abb65-6a13-4a80-9a28-89d32905384b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.911577] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319107, 'name': CreateVM_Task, 'duration_secs': 0.424182} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.912153] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1055.912839] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.913015] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.913360] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1055.913619] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63d905aa-f6c1-43dc-bc7a-024cc04b7f2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.919683] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1055.919683] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b64def-6a06-ea35-d3dc-4440ea24b141" [ 1055.919683] env[62521]: _type = "Task" [ 1055.919683] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.932172] env[62521]: DEBUG nova.compute.manager [req-c6cb6c0c-8f7f-43d1-904f-4d79322fb17a req-a37b07a4-dc87-436b-b122-5ae283a6b849 service nova] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Detach interface failed, port_id=dee03f5c-2442-438b-8786-6f3e0c35f8ab, reason: Instance 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1055.939289] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b64def-6a06-ea35-d3dc-4440ea24b141, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.961525] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5218f426-58f4-e753-5447-7e1e2c1f8e7c, 'name': SearchDatastore_Task, 'duration_secs': 0.013679} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.961811] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.962118] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] bc7d76b1-76b2-49a7-8188-0672312767a1/bc7d76b1-76b2-49a7-8188-0672312767a1.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1055.962394] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6121b8a5-017f-4b15-a050-05c3a91ce50e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.968955] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.970699] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.974046] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1055.974046] env[62521]: value = "task-1319108" [ 1055.974046] env[62521]: _type = "Task" [ 1055.974046] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.982393] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319108, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.133419] env[62521]: DEBUG nova.network.neutron [req-a942089e-fbc0-41fd-9b0b-25f6dd062bee req-fcfd2222-2b04-4246-84ac-3c9d7c25cf93 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updated VIF entry in instance network info cache for port 2beae869-6010-48ba-8e25-6b392dd2ec1f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1056.133419] env[62521]: DEBUG nova.network.neutron [req-a942089e-fbc0-41fd-9b0b-25f6dd062bee req-fcfd2222-2b04-4246-84ac-3c9d7c25cf93 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.165880] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319106, 'name': ReconfigVM_Task, 'duration_secs': 0.51966} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.166308] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 4021ce0d-69ee-4c57-bb18-fd77f339ed72/4021ce0d-69ee-4c57-bb18-fd77f339ed72.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1056.167526] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c1d7f676-a7af-4dc3-8686-c9de8091208e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.174944] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1056.174944] env[62521]: value = "task-1319109" [ 1056.174944] env[62521]: _type = "Task" [ 1056.174944] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.183929] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319109, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.187335] env[62521]: DEBUG nova.compute.utils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1056.188741] env[62521]: DEBUG nova.compute.manager [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1056.188918] env[62521]: DEBUG nova.network.neutron [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1056.235808] env[62521]: DEBUG nova.policy [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfab7121825d4e049370b526f90eb499', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11829634203c4967bbcff90eb904097c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1056.332568] env[62521]: INFO nova.compute.manager [-] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Took 1.49 seconds to deallocate network for instance. [ 1056.432733] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b64def-6a06-ea35-d3dc-4440ea24b141, 'name': SearchDatastore_Task, 'duration_secs': 0.022143} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.433176] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.433311] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1056.433560] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.433726] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.434148] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1056.434246] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-418c9061-5a05-4bab-9ec8-eeff02aa8241 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.442842] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1056.443036] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1056.443852] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90e2c364-bba1-47d5-9f28-ca438129f6df {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.453156] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1056.453156] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52efd2c3-2879-60d6-55e1-ecbcceaf4d96" [ 1056.453156] env[62521]: _type = "Task" [ 1056.453156] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.464585] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52efd2c3-2879-60d6-55e1-ecbcceaf4d96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.472340] env[62521]: DEBUG nova.compute.manager [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1056.485607] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319108, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.526807] env[62521]: DEBUG nova.network.neutron [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Successfully created port: c6383189-f73f-4f87-8424-e8956b34f94c {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1056.627765] env[62521]: DEBUG nova.network.neutron [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [{"id": "8824a231-01df-41e5-9bea-9071d8b41cb8", "address": "fa:16:3e:2e:cc:1b", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8824a231-01", "ovs_interfaceid": "8824a231-01df-41e5-9bea-9071d8b41cb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.638390] env[62521]: DEBUG oslo_concurrency.lockutils [req-a942089e-fbc0-41fd-9b0b-25f6dd062bee req-fcfd2222-2b04-4246-84ac-3c9d7c25cf93 service nova] Releasing lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.685175] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319109, 'name': Rename_Task, 'duration_secs': 0.326609} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.685476] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1056.685740] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34578660-f87d-451e-8a01-5b342b3d6dfa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.692053] env[62521]: DEBUG nova.compute.manager [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1056.699175] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1056.699175] env[62521]: value = "task-1319110" [ 1056.699175] env[62521]: _type = "Task" [ 1056.699175] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.706543] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319110, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.843858] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.843858] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.843858] env[62521]: DEBUG nova.objects.instance [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lazy-loading 'resources' on Instance uuid 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.961712] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52efd2c3-2879-60d6-55e1-ecbcceaf4d96, 'name': SearchDatastore_Task, 'duration_secs': 0.01535} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.962524] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a353953f-d518-409f-9372-2cc7d9b7bf35 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.967279] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1056.967279] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523a70fb-3acf-ad5b-118f-a13d96032a6b" [ 1056.967279] env[62521]: _type = "Task" [ 1056.967279] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.974637] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523a70fb-3acf-ad5b-118f-a13d96032a6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.987059] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319108, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.801769} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.988613] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] bc7d76b1-76b2-49a7-8188-0672312767a1/bc7d76b1-76b2-49a7-8188-0672312767a1.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1056.988875] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1056.989353] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4a23d63-37a3-4c9f-abba-c44802e6ba88 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.995549] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1056.995549] env[62521]: value = "task-1319111" [ 1056.995549] env[62521]: _type = "Task" [ 1056.995549] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.003276] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319111, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.004322] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.130867] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.131157] env[62521]: DEBUG nova.objects.instance [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'migration_context' on Instance uuid 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.210388] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319110, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.480463] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523a70fb-3acf-ad5b-118f-a13d96032a6b, 'name': SearchDatastore_Task, 'duration_secs': 0.029655} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.480773] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.480992] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 7cb9fe57-6940-4f84-bdde-32f185e28451/7cb9fe57-6940-4f84-bdde-32f185e28451.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1057.481276] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d5c69e6-00a7-4da2-b98e-8411a78a0222 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.488175] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1057.488175] env[62521]: value = "task-1319112" [ 1057.488175] env[62521]: _type = "Task" [ 1057.488175] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.498546] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319112, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.503893] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d854119d-c41e-487f-9024-353fe1b49dbb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.508929] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319111, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062885} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.509475] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1057.510223] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a0fcb8-7e11-4678-b029-cd4fb26ba8da {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.515099] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47aa2f9a-c6da-4643-a9de-a4973685ea22 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.535283] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] bc7d76b1-76b2-49a7-8188-0672312767a1/bc7d76b1-76b2-49a7-8188-0672312767a1.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1057.535927] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ab6edd9-1ce7-4b62-98f7-61abadb5a772 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.574318] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-640eb1e1-2050-47bf-8f3c-bc444aa20a1b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.582970] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefd8a6c-5bd6-4a95-8ecf-ddde293ac7d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.587117] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1057.587117] env[62521]: value = "task-1319113" [ 1057.587117] env[62521]: _type = "Task" [ 1057.587117] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.598063] env[62521]: DEBUG nova.compute.provider_tree [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.604660] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319113, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.634141] env[62521]: DEBUG nova.objects.base [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Object Instance<7b5afeab-f04e-453a-a3ed-09e3a725e8d6> lazy-loaded attributes: info_cache,migration_context {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1057.635091] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0a5de1-330b-441b-8831-87e000cf5d11 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.656193] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02c6144e-ac6b-4f88-bde0-fbbd8955085f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.662091] env[62521]: DEBUG oslo_vmware.api [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1057.662091] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a265f0-56bd-1f57-0b2e-e2f5ff58e741" [ 1057.662091] env[62521]: _type = "Task" [ 1057.662091] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.669702] env[62521]: DEBUG oslo_vmware.api [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a265f0-56bd-1f57-0b2e-e2f5ff58e741, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.705931] env[62521]: DEBUG nova.compute.manager [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1057.714136] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319110, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.734696] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1057.734985] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1057.735188] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.735384] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1057.735536] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.735691] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1057.735909] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1057.736086] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1057.736268] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1057.736449] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1057.736617] env[62521]: DEBUG nova.virt.hardware [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1057.737494] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bae1e11-6647-48c2-b63b-1b8e48f01083 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.745632] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27e8456-06a8-4ad7-86ef-7875ad5b60f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.988684] env[62521]: DEBUG nova.compute.manager [req-f6cfdbd6-b1e5-47c5-a4d8-cf4afb0b5dc9 req-78590204-eea3-411e-889d-45ba6b746a5b service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Received event network-vif-plugged-c6383189-f73f-4f87-8424-e8956b34f94c {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.989041] env[62521]: DEBUG oslo_concurrency.lockutils [req-f6cfdbd6-b1e5-47c5-a4d8-cf4afb0b5dc9 req-78590204-eea3-411e-889d-45ba6b746a5b service nova] Acquiring lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.989161] env[62521]: DEBUG oslo_concurrency.lockutils [req-f6cfdbd6-b1e5-47c5-a4d8-cf4afb0b5dc9 req-78590204-eea3-411e-889d-45ba6b746a5b service nova] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.989378] env[62521]: DEBUG oslo_concurrency.lockutils [req-f6cfdbd6-b1e5-47c5-a4d8-cf4afb0b5dc9 req-78590204-eea3-411e-889d-45ba6b746a5b service nova] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.989561] env[62521]: DEBUG nova.compute.manager [req-f6cfdbd6-b1e5-47c5-a4d8-cf4afb0b5dc9 req-78590204-eea3-411e-889d-45ba6b746a5b service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] No waiting events found dispatching network-vif-plugged-c6383189-f73f-4f87-8424-e8956b34f94c {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1057.989730] env[62521]: WARNING nova.compute.manager [req-f6cfdbd6-b1e5-47c5-a4d8-cf4afb0b5dc9 req-78590204-eea3-411e-889d-45ba6b746a5b service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Received unexpected event network-vif-plugged-c6383189-f73f-4f87-8424-e8956b34f94c for instance with vm_state building and task_state spawning. [ 1058.000022] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319112, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.087799] env[62521]: DEBUG nova.network.neutron [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Successfully updated port: c6383189-f73f-4f87-8424-e8956b34f94c {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1058.104666] env[62521]: DEBUG nova.scheduler.client.report [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.111165] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319113, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.173822] env[62521]: DEBUG oslo_vmware.api [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a265f0-56bd-1f57-0b2e-e2f5ff58e741, 'name': SearchDatastore_Task, 'duration_secs': 0.007612} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.174137] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.211071] env[62521]: DEBUG oslo_vmware.api [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319110, 'name': PowerOnVM_Task, 'duration_secs': 1.058649} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.211320] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1058.211547] env[62521]: INFO nova.compute.manager [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Took 10.04 seconds to spawn the instance on the hypervisor. [ 1058.211735] env[62521]: DEBUG nova.compute.manager [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1058.212538] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b651a20-29e9-4405-a531-f0e5c5d03854 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.508590] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319112, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.684261} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.509267] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 7cb9fe57-6940-4f84-bdde-32f185e28451/7cb9fe57-6940-4f84-bdde-32f185e28451.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1058.509502] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1058.509756] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44b8605a-dc4f-4bd7-8d6e-da0e20ad397f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.516697] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1058.516697] env[62521]: value = "task-1319114" [ 1058.516697] env[62521]: _type = "Task" [ 1058.516697] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.525123] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319114, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.595922] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.596030] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.596176] env[62521]: DEBUG nova.network.neutron [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1058.600419] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319113, 'name': ReconfigVM_Task, 'duration_secs': 0.657151} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.600936] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Reconfigured VM instance instance-00000064 to attach disk [datastore2] bc7d76b1-76b2-49a7-8188-0672312767a1/bc7d76b1-76b2-49a7-8188-0672312767a1.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1058.601591] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-18da985a-f33a-45f9-9d8e-f4276fc57955 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.608311] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1058.608311] env[62521]: value = "task-1319115" [ 1058.608311] env[62521]: _type = "Task" [ 1058.608311] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.613833] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.619432] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.615s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.621065] env[62521]: INFO nova.compute.claims [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1058.623897] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319115, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.638932] env[62521]: INFO nova.scheduler.client.report [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Deleted allocations for instance 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c [ 1058.730045] env[62521]: INFO nova.compute.manager [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Took 17.70 seconds to build instance. [ 1058.920795] env[62521]: DEBUG oslo_vmware.rw_handles [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521ce241-c3ea-9fbe-98e3-05b5cb802e8a/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1058.921862] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5551fa-842a-4215-8bb0-43d97e00ff2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.928428] env[62521]: DEBUG oslo_vmware.rw_handles [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521ce241-c3ea-9fbe-98e3-05b5cb802e8a/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1058.928611] env[62521]: ERROR oslo_vmware.rw_handles [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521ce241-c3ea-9fbe-98e3-05b5cb802e8a/disk-0.vmdk due to incomplete transfer. [ 1058.928844] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-36e1f3f8-53ec-4364-82e5-9226814af45c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.936590] env[62521]: DEBUG oslo_vmware.rw_handles [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521ce241-c3ea-9fbe-98e3-05b5cb802e8a/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1058.936974] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Uploaded image 9d0ca44a-c290-4856-a6d8-d11e3d6b1026 to the Glance image server {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1058.939368] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1058.939554] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fd64f86a-2f46-4bd1-bac0-94e13362c4e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.947930] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1058.947930] env[62521]: value = "task-1319116" [ 1058.947930] env[62521]: _type = "Task" [ 1058.947930] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.956743] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319116, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.026698] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319114, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104328} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.027062] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1059.027616] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47f24a26-b324-471b-b6f8-ad76e280364a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.049383] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 7cb9fe57-6940-4f84-bdde-32f185e28451/7cb9fe57-6940-4f84-bdde-32f185e28451.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.049689] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07bb1321-d1c9-4fd1-80c7-f17d2f749257 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.068593] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1059.068593] env[62521]: value = "task-1319117" [ 1059.068593] env[62521]: _type = "Task" [ 1059.068593] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.076478] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319117, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.119311] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319115, 'name': Rename_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.141038] env[62521]: DEBUG nova.network.neutron [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1059.147724] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6d2c124e-5bdd-4111-976f-a7dae548d6d9 tempest-AttachVolumeTestJSON-1556758036 tempest-AttachVolumeTestJSON-1556758036-project-member] Lock "7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.463s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.233890] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ae393fab-c3b6-47ef-9106-6ae17f99464d tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.209s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.315156] env[62521]: DEBUG nova.network.neutron [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance_info_cache with network_info: [{"id": "c6383189-f73f-4f87-8424-e8956b34f94c", "address": "fa:16:3e:ba:1a:11", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6383189-f7", "ovs_interfaceid": "c6383189-f73f-4f87-8424-e8956b34f94c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.359683] env[62521]: INFO nova.compute.manager [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Rescuing [ 1059.359963] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "refresh_cache-4021ce0d-69ee-4c57-bb18-fd77f339ed72" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.360142] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "refresh_cache-4021ce0d-69ee-4c57-bb18-fd77f339ed72" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.360362] env[62521]: DEBUG nova.network.neutron [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1059.457804] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319116, 'name': Destroy_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.578987] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.619114] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319115, 'name': Rename_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.762184] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ebbf51-7da2-408b-8cec-384ac89228b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.769785] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5df2c01-952f-4f44-b7cf-5245d2eb2484 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.798919] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec2ba37-b1cb-4ad7-87f7-adca2ce02668 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.806326] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6066677b-1d71-4570-b3dd-f52ecdc732a8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.819453] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.819747] env[62521]: DEBUG nova.compute.manager [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Instance network_info: |[{"id": "c6383189-f73f-4f87-8424-e8956b34f94c", "address": "fa:16:3e:ba:1a:11", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6383189-f7", "ovs_interfaceid": "c6383189-f73f-4f87-8424-e8956b34f94c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1059.820212] env[62521]: DEBUG nova.compute.provider_tree [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.821586] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:1a:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6383189-f73f-4f87-8424-e8956b34f94c', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.828990] env[62521]: DEBUG oslo.service.loopingcall [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.829922] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1059.830154] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8d57d96-e6b9-47fc-8371-f726b4cd5495 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.849955] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.849955] env[62521]: value = "task-1319119" [ 1059.849955] env[62521]: _type = "Task" [ 1059.849955] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.856962] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319119, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.893782] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.893983] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.894148] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1059.957948] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319116, 'name': Destroy_Task, 'duration_secs': 0.89639} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.958175] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Destroyed the VM [ 1059.958390] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1059.958718] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a34f0ece-58d5-461d-8542-4e3deef7929b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.965926] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1059.965926] env[62521]: value = "task-1319120" [ 1059.965926] env[62521]: _type = "Task" [ 1059.965926] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.974873] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319120, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.033385] env[62521]: DEBUG nova.compute.manager [req-1678cd26-0f4e-4275-a88b-9529c23073a7 req-dfaba728-f7ac-46ad-bf0c-9cbf30774de5 service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Received event network-changed-c6383189-f73f-4f87-8424-e8956b34f94c {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.033649] env[62521]: DEBUG nova.compute.manager [req-1678cd26-0f4e-4275-a88b-9529c23073a7 req-dfaba728-f7ac-46ad-bf0c-9cbf30774de5 service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Refreshing instance network info cache due to event network-changed-c6383189-f73f-4f87-8424-e8956b34f94c. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1060.033807] env[62521]: DEBUG oslo_concurrency.lockutils [req-1678cd26-0f4e-4275-a88b-9529c23073a7 req-dfaba728-f7ac-46ad-bf0c-9cbf30774de5 service nova] Acquiring lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.033953] env[62521]: DEBUG oslo_concurrency.lockutils [req-1678cd26-0f4e-4275-a88b-9529c23073a7 req-dfaba728-f7ac-46ad-bf0c-9cbf30774de5 service nova] Acquired lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.034139] env[62521]: DEBUG nova.network.neutron [req-1678cd26-0f4e-4275-a88b-9529c23073a7 req-dfaba728-f7ac-46ad-bf0c-9cbf30774de5 service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Refreshing network info cache for port c6383189-f73f-4f87-8424-e8956b34f94c {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1060.065050] env[62521]: DEBUG nova.network.neutron [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Updating instance_info_cache with network_info: [{"id": "59e6ef56-abf3-4b61-ab38-86918b4f4ce1", "address": "fa:16:3e:59:05:3e", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59e6ef56-ab", "ovs_interfaceid": "59e6ef56-abf3-4b61-ab38-86918b4f4ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.078600] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.121205] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319115, 'name': Rename_Task, 'duration_secs': 1.326491} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.121521] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1060.121806] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9e91294-236b-4aac-a389-2e54f120caec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.128533] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1060.128533] env[62521]: value = "task-1319121" [ 1060.128533] env[62521]: _type = "Task" [ 1060.128533] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.136503] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319121, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.330933] env[62521]: DEBUG nova.scheduler.client.report [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1060.361939] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319119, 'name': CreateVM_Task, 'duration_secs': 0.475485} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.362171] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1060.362965] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.363197] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.363570] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1060.363851] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24a70c37-94d1-4ad7-80f8-9bb4c03381ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.369860] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1060.369860] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d0c405-45fe-506e-2004-14d47b2dcaa2" [ 1060.369860] env[62521]: _type = "Task" [ 1060.369860] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.377942] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d0c405-45fe-506e-2004-14d47b2dcaa2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.476607] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319120, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.569503] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "refresh_cache-4021ce0d-69ee-4c57-bb18-fd77f339ed72" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.580904] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319117, 'name': ReconfigVM_Task, 'duration_secs': 1.300826} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.581238] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 7cb9fe57-6940-4f84-bdde-32f185e28451/7cb9fe57-6940-4f84-bdde-32f185e28451.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.581893] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b91fd5d6-907a-485f-aea7-0820d276bd8b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.588681] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1060.588681] env[62521]: value = "task-1319122" [ 1060.588681] env[62521]: _type = "Task" [ 1060.588681] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.600287] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319122, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.640262] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319121, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.779046] env[62521]: DEBUG nova.network.neutron [req-1678cd26-0f4e-4275-a88b-9529c23073a7 req-dfaba728-f7ac-46ad-bf0c-9cbf30774de5 service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updated VIF entry in instance network info cache for port c6383189-f73f-4f87-8424-e8956b34f94c. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1060.779541] env[62521]: DEBUG nova.network.neutron [req-1678cd26-0f4e-4275-a88b-9529c23073a7 req-dfaba728-f7ac-46ad-bf0c-9cbf30774de5 service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance_info_cache with network_info: [{"id": "c6383189-f73f-4f87-8424-e8956b34f94c", "address": "fa:16:3e:ba:1a:11", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6383189-f7", "ovs_interfaceid": "c6383189-f73f-4f87-8424-e8956b34f94c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.837605] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.218s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.840060] env[62521]: DEBUG nova.compute.manager [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1060.842204] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.667s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.880232] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52d0c405-45fe-506e-2004-14d47b2dcaa2, 'name': SearchDatastore_Task, 'duration_secs': 0.011237} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.880556] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.881063] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.881063] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.881225] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.881478] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.882258] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49409bf4-76dc-4848-980f-2a9913ab6120 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.892017] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.892223] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1060.892942] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e14296e-faaa-46cf-b8a0-1331d8ab6b1d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.898289] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1060.898289] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527d52b2-38aa-4d15-f102-ca86a4ee4c7f" [ 1060.898289] env[62521]: _type = "Task" [ 1060.898289] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.907162] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527d52b2-38aa-4d15-f102-ca86a4ee4c7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.976393] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319120, 'name': RemoveSnapshot_Task, 'duration_secs': 0.588418} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.976725] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1060.977025] env[62521]: DEBUG nova.compute.manager [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1060.977965] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eea3f37-c9f4-4dbf-9147-bacde0eb8ffe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.098644] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319122, 'name': Rename_Task, 'duration_secs': 0.274838} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.099081] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1061.099433] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e23d378e-1728-4470-9480-3d183c10078d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.109413] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1061.109413] env[62521]: value = "task-1319123" [ 1061.109413] env[62521]: _type = "Task" [ 1061.109413] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.110068] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.110453] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9cb97620-6e5c-4873-8901-6ba004e6ba6a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.122785] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319123, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.124280] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1061.124280] env[62521]: value = "task-1319124" [ 1061.124280] env[62521]: _type = "Task" [ 1061.124280] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.132920] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319124, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.142056] env[62521]: DEBUG oslo_vmware.api [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319121, 'name': PowerOnVM_Task, 'duration_secs': 0.732668} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.142366] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1061.142621] env[62521]: INFO nova.compute.manager [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Took 10.69 seconds to spawn the instance on the hypervisor. [ 1061.142887] env[62521]: DEBUG nova.compute.manager [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1061.143652] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54e78ff-1b0f-4565-b5a7-f85f31276aaf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.282812] env[62521]: DEBUG oslo_concurrency.lockutils [req-1678cd26-0f4e-4275-a88b-9529c23073a7 req-dfaba728-f7ac-46ad-bf0c-9cbf30774de5 service nova] Releasing lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.345319] env[62521]: DEBUG nova.compute.utils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1061.350421] env[62521]: DEBUG nova.compute.manager [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1061.350421] env[62521]: DEBUG nova.network.neutron [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1061.406071] env[62521]: DEBUG nova.policy [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba9b0b82d53e4502bef18a9f031fcc9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5780fc8ecfd64161bc2ea0b3a092b4d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1061.419057] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]527d52b2-38aa-4d15-f102-ca86a4ee4c7f, 'name': SearchDatastore_Task, 'duration_secs': 0.02827} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.419949] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dde9d679-9aa5-4cf5-acca-c5d634ba488c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.425935] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1061.425935] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c65ca5-a394-d4d5-b679-d72a3a6038f6" [ 1061.425935] env[62521]: _type = "Task" [ 1061.425935] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.438054] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c65ca5-a394-d4d5-b679-d72a3a6038f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.460965] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.460965] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquired lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.460965] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Forcefully refreshing network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1061.492554] env[62521]: INFO nova.compute.manager [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Shelve offloading [ 1061.494572] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.494780] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c91daf53-d1b0-418b-a10a-d1885e156221 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.501983] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1061.501983] env[62521]: value = "task-1319126" [ 1061.501983] env[62521]: _type = "Task" [ 1061.501983] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.514881] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c901e6-b00b-4dcb-8517-0030a264f5e8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.521853] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1061.522029] env[62521]: DEBUG nova.compute.manager [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1061.522957] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd1703c5-2239-4df1-86c2-b7efa2e6bb08 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.528303] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd93a2a4-e28f-45b5-9e17-2e9ad76b9f44 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.532837] env[62521]: DEBUG oslo_concurrency.lockutils [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.533026] env[62521]: DEBUG oslo_concurrency.lockutils [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.533241] env[62521]: DEBUG nova.network.neutron [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1061.561445] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45eeb627-46cc-4481-bbb7-7db0eb093068 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.569243] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890e0ae1-5d75-4522-b088-e77313c5b656 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.583448] env[62521]: DEBUG nova.compute.provider_tree [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.621603] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319123, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.633115] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319124, 'name': PowerOffVM_Task, 'duration_secs': 0.233579} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.633396] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1061.634204] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29645195-340b-4aaf-a308-94550010d167 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.652976] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63597557-33bf-4b19-8cc2-dc58c1c65526 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.666398] env[62521]: INFO nova.compute.manager [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Took 17.77 seconds to build instance. [ 1061.698856] env[62521]: DEBUG nova.network.neutron [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Successfully created port: f4903211-c60f-4fa2-9589-a98babc4b5c5 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1061.707608] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.707946] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7173d45c-81d5-455b-aded-259d5c7fc739 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.720651] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1061.720651] env[62521]: value = "task-1319127" [ 1061.720651] env[62521]: _type = "Task" [ 1061.720651] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.728639] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1061.728863] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.729131] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.729351] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.729563] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.729813] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-779f21cc-fba2-4e76-b14b-5501004d3b39 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.740534] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1061.740744] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1061.742188] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34c1835f-d2ee-4fce-8f14-d693bbc66d3f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.746832] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1061.746832] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c2ae00-f3aa-668e-ba94-bae83c80ac7c" [ 1061.746832] env[62521]: _type = "Task" [ 1061.746832] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.754348] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c2ae00-f3aa-668e-ba94-bae83c80ac7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.852774] env[62521]: DEBUG nova.compute.manager [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1061.941234] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c65ca5-a394-d4d5-b679-d72a3a6038f6, 'name': SearchDatastore_Task, 'duration_secs': 0.013422} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.941741] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.942651] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 722aa42e-f99c-42a6-aebc-f5ea99447e2f/722aa42e-f99c-42a6-aebc-f5ea99447e2f.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1061.942651] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39a4f3e1-583e-4695-946e-13235d457812 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.949539] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1061.949539] env[62521]: value = "task-1319128" [ 1061.949539] env[62521]: _type = "Task" [ 1061.949539] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.958487] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319128, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.088863] env[62521]: DEBUG nova.scheduler.client.report [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1062.125856] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319123, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.171848] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e6f8fd1-0996-49a1-bd84-da603d6dc9a1 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "bc7d76b1-76b2-49a7-8188-0672312767a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.282s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.267262] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52c2ae00-f3aa-668e-ba94-bae83c80ac7c, 'name': SearchDatastore_Task, 'duration_secs': 0.017248} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.271017] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fa5d2d8-7730-4b1c-bcc3-b07dd95babc6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.277753] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1062.277753] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523a5bfd-52cb-e83f-1ced-f0546a019f90" [ 1062.277753] env[62521]: _type = "Task" [ 1062.277753] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.290719] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523a5bfd-52cb-e83f-1ced-f0546a019f90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.361013] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.361296] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.361478] env[62521]: INFO nova.compute.manager [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Shelving [ 1062.461175] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319128, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478098} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.461538] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 722aa42e-f99c-42a6-aebc-f5ea99447e2f/722aa42e-f99c-42a6-aebc-f5ea99447e2f.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1062.461766] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1062.462034] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2538b140-2c32-41dd-92ca-aab272ac6362 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.467952] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1062.467952] env[62521]: value = "task-1319129" [ 1062.467952] env[62521]: _type = "Task" [ 1062.467952] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.475687] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319129, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.501989] env[62521]: DEBUG nova.network.neutron [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Updating instance_info_cache with network_info: [{"id": "9ee94c09-d3e5-465a-9926-639300512161", "address": "fa:16:3e:0b:5a:3b", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ee94c09-d3", "ovs_interfaceid": "9ee94c09-d3e5-465a-9926-639300512161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.624762] env[62521]: DEBUG oslo_vmware.api [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319123, 'name': PowerOnVM_Task, 'duration_secs': 1.453388} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.625054] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1062.625264] env[62521]: INFO nova.compute.manager [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Took 9.93 seconds to spawn the instance on the hypervisor. [ 1062.625450] env[62521]: DEBUG nova.compute.manager [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1062.626233] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00c2f78-ea5e-4edb-b8a3-d314e3475df0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.788065] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523a5bfd-52cb-e83f-1ced-f0546a019f90, 'name': SearchDatastore_Task, 'duration_secs': 0.017017} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.788425] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.788691] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 4021ce0d-69ee-4c57-bb18-fd77f339ed72/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk. {{(pid=62521) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1062.788948] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc1f4eec-dfc4-49d1-9fea-e011799bb805 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.795056] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1062.795056] env[62521]: value = "task-1319130" [ 1062.795056] env[62521]: _type = "Task" [ 1062.795056] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.802521] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.803370] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [{"id": "8824a231-01df-41e5-9bea-9071d8b41cb8", "address": "fa:16:3e:2e:cc:1b", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8824a231-01", "ovs_interfaceid": "8824a231-01df-41e5-9bea-9071d8b41cb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.864040] env[62521]: DEBUG nova.compute.manager [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1062.872627] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1062.872899] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2dedda6-3f71-4fa5-99f8-78a174e191c6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.879683] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1062.879683] env[62521]: value = "task-1319131" [ 1062.879683] env[62521]: _type = "Task" [ 1062.879683] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.888170] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319131, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.891508] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.891742] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.891903] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.892106] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.892263] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.892414] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.892652] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.892790] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.892957] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.893137] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.893314] env[62521]: DEBUG nova.virt.hardware [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.894110] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7775b1-60b3-47f8-b364-e20af6aa34e8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.901052] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee86faf-816d-48ad-86ea-d7412f8eabd1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.977706] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319129, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06973} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.977988] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1062.978824] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356fdeb6-aaa5-4ca7-a75d-4385e5a8d099 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.001177] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 722aa42e-f99c-42a6-aebc-f5ea99447e2f/722aa42e-f99c-42a6-aebc-f5ea99447e2f.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.001545] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bb33d39-3d3b-4514-926c-735ea49b6ec9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.016596] env[62521]: DEBUG oslo_concurrency.lockutils [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.024812] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1063.024812] env[62521]: value = "task-1319132" [ 1063.024812] env[62521]: _type = "Task" [ 1063.024812] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.033660] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319132, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.102236] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.259s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.112021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "bc7d76b1-76b2-49a7-8188-0672312767a1" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.112021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "bc7d76b1-76b2-49a7-8188-0672312767a1" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.112021] env[62521]: DEBUG nova.compute.manager [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1063.112021] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99203bea-b9ac-463b-9ed1-c85f50ec80e4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.118548] env[62521]: DEBUG nova.compute.manager [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62521) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1063.119623] env[62521]: DEBUG nova.objects.instance [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lazy-loading 'flavor' on Instance uuid bc7d76b1-76b2-49a7-8188-0672312767a1 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.126057] env[62521]: DEBUG nova.compute.manager [req-d6af106d-8beb-4be5-81f4-e361543ae68c req-13933b40-04e2-4f3e-a999-bb378fae7da1 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Received event network-vif-plugged-f4903211-c60f-4fa2-9589-a98babc4b5c5 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.127463] env[62521]: DEBUG oslo_concurrency.lockutils [req-d6af106d-8beb-4be5-81f4-e361543ae68c req-13933b40-04e2-4f3e-a999-bb378fae7da1 service nova] Acquiring lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.127826] env[62521]: DEBUG oslo_concurrency.lockutils [req-d6af106d-8beb-4be5-81f4-e361543ae68c req-13933b40-04e2-4f3e-a999-bb378fae7da1 service nova] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.128144] env[62521]: DEBUG oslo_concurrency.lockutils [req-d6af106d-8beb-4be5-81f4-e361543ae68c req-13933b40-04e2-4f3e-a999-bb378fae7da1 service nova] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.128493] env[62521]: DEBUG nova.compute.manager [req-d6af106d-8beb-4be5-81f4-e361543ae68c req-13933b40-04e2-4f3e-a999-bb378fae7da1 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] No waiting events found dispatching network-vif-plugged-f4903211-c60f-4fa2-9589-a98babc4b5c5 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1063.128808] env[62521]: WARNING nova.compute.manager [req-d6af106d-8beb-4be5-81f4-e361543ae68c req-13933b40-04e2-4f3e-a999-bb378fae7da1 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Received unexpected event network-vif-plugged-f4903211-c60f-4fa2-9589-a98babc4b5c5 for instance with vm_state building and task_state spawning. [ 1063.148952] env[62521]: INFO nova.compute.manager [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Took 18.98 seconds to build instance. [ 1063.249068] env[62521]: DEBUG nova.network.neutron [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Successfully updated port: f4903211-c60f-4fa2-9589-a98babc4b5c5 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1063.306103] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Releasing lock "refresh_cache-7b5afeab-f04e-453a-a3ed-09e3a725e8d6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.306351] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updated the network info_cache for instance {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1063.306606] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319130, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.306822] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1063.306983] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1063.307150] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1063.307302] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1063.307452] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1063.307597] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1063.307728] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1063.307871] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1063.390014] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319131, 'name': PowerOffVM_Task, 'duration_secs': 0.167966} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.390322] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1063.391110] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ee4322-d080-4446-b8af-0ba467f2be65 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.414642] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4992cc5a-6d4c-4b7c-94b6-be71139d23b6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.467770] env[62521]: DEBUG nova.compute.manager [req-630413fa-4f06-487f-8f25-8652afa3d33b req-eb19fd3c-9e43-48cc-bc87-939df5d5a4ca service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Received event network-vif-unplugged-9ee94c09-d3e5-465a-9926-639300512161 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.468183] env[62521]: DEBUG oslo_concurrency.lockutils [req-630413fa-4f06-487f-8f25-8652afa3d33b req-eb19fd3c-9e43-48cc-bc87-939df5d5a4ca service nova] Acquiring lock "8b4db43f-bc21-4d40-9410-f643971e53fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.468573] env[62521]: DEBUG oslo_concurrency.lockutils [req-630413fa-4f06-487f-8f25-8652afa3d33b req-eb19fd3c-9e43-48cc-bc87-939df5d5a4ca service nova] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.468882] env[62521]: DEBUG oslo_concurrency.lockutils [req-630413fa-4f06-487f-8f25-8652afa3d33b req-eb19fd3c-9e43-48cc-bc87-939df5d5a4ca service nova] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.469200] env[62521]: DEBUG nova.compute.manager [req-630413fa-4f06-487f-8f25-8652afa3d33b req-eb19fd3c-9e43-48cc-bc87-939df5d5a4ca service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] No waiting events found dispatching network-vif-unplugged-9ee94c09-d3e5-465a-9926-639300512161 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1063.469548] env[62521]: WARNING nova.compute.manager [req-630413fa-4f06-487f-8f25-8652afa3d33b req-eb19fd3c-9e43-48cc-bc87-939df5d5a4ca service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Received unexpected event network-vif-unplugged-9ee94c09-d3e5-465a-9926-639300512161 for instance with vm_state shelved and task_state shelving_offloading. [ 1063.492074] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1063.492074] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907831fa-ecf4-4c34-aff6-cd9424ec1f43 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.501105] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1063.501392] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-86c92745-97a7-4ff7-8db4-90f56cfb884d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.535619] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319132, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.567557] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1063.567797] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1063.567993] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleting the datastore file [datastore2] 8b4db43f-bc21-4d40-9410-f643971e53fc {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1063.568339] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b79db96f-a07d-4a5f-aca2-1617927d88d4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.574387] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1063.574387] env[62521]: value = "task-1319134" [ 1063.574387] env[62521]: _type = "Task" [ 1063.574387] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.583696] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319134, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.626303] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1063.626578] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b281840-d26d-4d42-a5f8-e440313e6465 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.634968] env[62521]: DEBUG oslo_vmware.api [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1063.634968] env[62521]: value = "task-1319136" [ 1063.634968] env[62521]: _type = "Task" [ 1063.634968] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.642827] env[62521]: DEBUG oslo_vmware.api [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319136, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.656791] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4176e0c3-f352-4e22-9d6b-0c433f58d24d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.496s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.672019] env[62521]: INFO nova.scheduler.client.report [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleted allocation for migration ff0bde79-bf70-463f-ae4e-b1a86e2a6db9 [ 1063.752023] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "refresh_cache-3aa60da8-b251-4f31-bf8d-a4cd143b491e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.752023] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "refresh_cache-3aa60da8-b251-4f31-bf8d-a4cd143b491e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.752157] env[62521]: DEBUG nova.network.neutron [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1063.805994] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319130, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55539} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.806313] env[62521]: INFO nova.virt.vmwareapi.ds_util [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 4021ce0d-69ee-4c57-bb18-fd77f339ed72/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk. [ 1063.807715] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcfc0144-2540-4e11-8a3c-5c5549652536 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.810777] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.810976] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.811259] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.811326] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1063.812062] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71190797-1231-45ec-ae84-5d235613c64f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.836613] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 4021ce0d-69ee-4c57-bb18-fd77f339ed72/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.837372] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47feceb6-a13c-4c1f-a606-a4bfdb954735 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.854560] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85e12c9-4079-4e67-8f1d-2b270300431d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.859383] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1063.859383] env[62521]: value = "task-1319137" [ 1063.859383] env[62521]: _type = "Task" [ 1063.859383] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.874052] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1692fab8-dc4f-4ad9-b1fa-ca25790db1e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.881514] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.886436] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f561121-6c1b-4d8d-b11f-0bb1db45e048 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.917052] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179841MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1063.917220] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.917401] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.925378] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1063.925686] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f1048fbd-5e7b-422e-b871-621a5a7808a9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.933734] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1063.933734] env[62521]: value = "task-1319138" [ 1063.933734] env[62521]: _type = "Task" [ 1063.933734] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.941998] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319138, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.035147] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319132, 'name': ReconfigVM_Task, 'duration_secs': 0.609885} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.035454] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 722aa42e-f99c-42a6-aebc-f5ea99447e2f/722aa42e-f99c-42a6-aebc-f5ea99447e2f.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.036190] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e957e0ba-c766-4a23-a9d2-4fd8588c90c4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.042231] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1064.042231] env[62521]: value = "task-1319139" [ 1064.042231] env[62521]: _type = "Task" [ 1064.042231] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.049995] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319139, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.083752] env[62521]: DEBUG oslo_vmware.api [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319134, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152261} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.084037] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1064.084217] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1064.084399] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1064.104817] env[62521]: INFO nova.scheduler.client.report [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted allocations for instance 8b4db43f-bc21-4d40-9410-f643971e53fc [ 1064.148065] env[62521]: DEBUG oslo_vmware.api [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319136, 'name': PowerOffVM_Task, 'duration_secs': 0.439592} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.148637] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1064.148783] env[62521]: DEBUG nova.compute.manager [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1064.151193] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79040472-c49c-4821-87e6-9e98381e32d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.178385] env[62521]: DEBUG oslo_concurrency.lockutils [None req-68ad9893-5fef-4fc3-8ed4-1e60e598f691 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.428s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.297674] env[62521]: DEBUG nova.network.neutron [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1064.372167] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.444595] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319138, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.515836] env[62521]: DEBUG nova.network.neutron [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Updating instance_info_cache with network_info: [{"id": "f4903211-c60f-4fa2-9589-a98babc4b5c5", "address": "fa:16:3e:a2:d7:bd", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4903211-c6", "ovs_interfaceid": "f4903211-c60f-4fa2-9589-a98babc4b5c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.552305] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319139, 'name': Rename_Task, 'duration_secs': 0.243205} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.552605] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.552861] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2779d3ac-45b7-4b7e-adbc-ff4620e35fb6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.559069] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1064.559069] env[62521]: value = "task-1319140" [ 1064.559069] env[62521]: _type = "Task" [ 1064.559069] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.566563] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319140, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.609526] env[62521]: DEBUG oslo_concurrency.lockutils [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.663246] env[62521]: DEBUG oslo_concurrency.lockutils [None req-73c40082-0ced-4837-b8ce-6300ec412404 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "bc7d76b1-76b2-49a7-8188-0672312767a1" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.554s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.869956] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319137, 'name': ReconfigVM_Task, 'duration_secs': 0.831427} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.870278] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 4021ce0d-69ee-4c57-bb18-fd77f339ed72/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.871133] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c312a7f-702b-4f49-b37a-856f9b3f5f81 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.891746] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.892018] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.892240] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.892430] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.892608] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.899667] env[62521]: INFO nova.compute.manager [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Terminating instance [ 1064.901243] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b033af33-749c-4e79-9b93-891b7840a74b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.911831] env[62521]: DEBUG nova.compute.manager [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1064.912046] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1064.912864] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e8dea7-ebbd-41d6-b9ff-d468a710bda4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.920236] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1064.921432] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afd553b7-c90e-44bb-be62-56c8d178508a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.922931] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1064.922931] env[62521]: value = "task-1319141" [ 1064.922931] env[62521]: _type = "Task" [ 1064.922931] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.927104] env[62521]: DEBUG oslo_vmware.api [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1064.927104] env[62521]: value = "task-1319142" [ 1064.927104] env[62521]: _type = "Task" [ 1064.927104] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.937617] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319141, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.947113] env[62521]: DEBUG oslo_vmware.api [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.950945] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319138, 'name': CreateSnapshot_Task, 'duration_secs': 0.943107} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.951313] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1064.952352] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2e6bbf-4a12-4305-8ef9-5586ab53da92 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.955874] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance d574f077-b221-4f91-8b54-0915421cb36f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.956047] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.956216] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.956367] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 703ff423-98b1-4942-b6fd-8b95fe57bd0a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.956498] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 4021ce0d-69ee-4c57-bb18-fd77f339ed72 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.956629] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance bc7d76b1-76b2-49a7-8188-0672312767a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.956751] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 7cb9fe57-6940-4f84-bdde-32f185e28451 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.956862] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 722aa42e-f99c-42a6-aebc-f5ea99447e2f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.957029] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 3aa60da8-b251-4f31-bf8d-a4cd143b491e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.957319] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1064.957477] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1065.019734] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "refresh_cache-3aa60da8-b251-4f31-bf8d-a4cd143b491e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.020358] env[62521]: DEBUG nova.compute.manager [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Instance network_info: |[{"id": "f4903211-c60f-4fa2-9589-a98babc4b5c5", "address": "fa:16:3e:a2:d7:bd", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4903211-c6", "ovs_interfaceid": "f4903211-c60f-4fa2-9589-a98babc4b5c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1065.020800] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:d7:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd998416-f3d6-4a62-b828-5011063ce76a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f4903211-c60f-4fa2-9589-a98babc4b5c5', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1065.032710] env[62521]: DEBUG oslo.service.loopingcall [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1065.036815] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1065.037735] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b35623b1-bbf6-4e9f-8e11-db0078095c9b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.076353] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319140, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.080375] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1065.080375] env[62521]: value = "task-1319143" [ 1065.080375] env[62521]: _type = "Task" [ 1065.080375] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.089618] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319143, 'name': CreateVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.117666] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f412c2d8-061f-4c46-81c7-a1ab77e8c192 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.126713] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3706e192-ef6f-48ae-b8e3-a10e47c7dc89 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.163382] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ce181d-1ee5-4eef-8164-61cae454bf2a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.173186] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2729ce0-afaa-4c0f-bc17-8de35bed39ea {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.181105] env[62521]: DEBUG nova.compute.manager [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Received event network-changed-f4903211-c60f-4fa2-9589-a98babc4b5c5 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.181338] env[62521]: DEBUG nova.compute.manager [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Refreshing instance network info cache due to event network-changed-f4903211-c60f-4fa2-9589-a98babc4b5c5. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1065.181584] env[62521]: DEBUG oslo_concurrency.lockutils [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] Acquiring lock "refresh_cache-3aa60da8-b251-4f31-bf8d-a4cd143b491e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.181734] env[62521]: DEBUG oslo_concurrency.lockutils [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] Acquired lock "refresh_cache-3aa60da8-b251-4f31-bf8d-a4cd143b491e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.181912] env[62521]: DEBUG nova.network.neutron [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Refreshing network info cache for port f4903211-c60f-4fa2-9589-a98babc4b5c5 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1065.196574] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.434892] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319141, 'name': ReconfigVM_Task, 'duration_secs': 0.26395} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.435613] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1065.435888] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4af5cb9d-ab9e-49b9-9d58-961f8c46ba41 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.440529] env[62521]: DEBUG oslo_vmware.api [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319142, 'name': PowerOffVM_Task, 'duration_secs': 0.212056} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.441086] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1065.441269] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1065.441512] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1af9956e-6b7e-4722-b349-82e9be055d1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.445347] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1065.445347] env[62521]: value = "task-1319144" [ 1065.445347] env[62521]: _type = "Task" [ 1065.445347] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.452414] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319144, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.483929] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1065.484273] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9984f91a-c16d-4962-9b2e-19a46cc4cf38 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.493241] env[62521]: DEBUG nova.compute.manager [req-1f86e073-d523-4f84-99a5-1d3357649f4d req-6394b70b-2bbf-4cf0-a06e-e82891d21127 service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Received event network-changed-9ee94c09-d3e5-465a-9926-639300512161 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.493467] env[62521]: DEBUG nova.compute.manager [req-1f86e073-d523-4f84-99a5-1d3357649f4d req-6394b70b-2bbf-4cf0-a06e-e82891d21127 service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Refreshing instance network info cache due to event network-changed-9ee94c09-d3e5-465a-9926-639300512161. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1065.493713] env[62521]: DEBUG oslo_concurrency.lockutils [req-1f86e073-d523-4f84-99a5-1d3357649f4d req-6394b70b-2bbf-4cf0-a06e-e82891d21127 service nova] Acquiring lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.493875] env[62521]: DEBUG oslo_concurrency.lockutils [req-1f86e073-d523-4f84-99a5-1d3357649f4d req-6394b70b-2bbf-4cf0-a06e-e82891d21127 service nova] Acquired lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.494114] env[62521]: DEBUG nova.network.neutron [req-1f86e073-d523-4f84-99a5-1d3357649f4d req-6394b70b-2bbf-4cf0-a06e-e82891d21127 service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Refreshing network info cache for port 9ee94c09-d3e5-465a-9926-639300512161 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1065.496838] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1065.496838] env[62521]: value = "task-1319146" [ 1065.496838] env[62521]: _type = "Task" [ 1065.496838] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.506029] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319146, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.533537] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1065.533806] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1065.534007] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleting the datastore file [datastore1] 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1065.534308] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1967dbe-d392-4a4c-9479-a26daaa399d2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.541637] env[62521]: DEBUG oslo_vmware.api [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1065.541637] env[62521]: value = "task-1319147" [ 1065.541637] env[62521]: _type = "Task" [ 1065.541637] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.549556] env[62521]: DEBUG oslo_vmware.api [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319147, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.576504] env[62521]: DEBUG oslo_vmware.api [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319140, 'name': PowerOnVM_Task, 'duration_secs': 0.717982} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.577475] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.577475] env[62521]: INFO nova.compute.manager [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Took 7.87 seconds to spawn the instance on the hypervisor. [ 1065.577475] env[62521]: DEBUG nova.compute.manager [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.578132] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50559dda-95da-4ee6-aa13-13f6e404fd16 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.596230] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319143, 'name': CreateVM_Task, 'duration_secs': 0.408364} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.596459] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1065.597199] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.597348] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.597696] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1065.597950] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11451af0-2f51-4a46-b873-114747f373ae {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.602913] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1065.602913] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5264cd02-ab4e-c3d3-f1aa-76436c5bb8a4" [ 1065.602913] env[62521]: _type = "Task" [ 1065.602913] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.610950] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5264cd02-ab4e-c3d3-f1aa-76436c5bb8a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.688474] env[62521]: DEBUG oslo_concurrency.lockutils [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "8b4db43f-bc21-4d40-9410-f643971e53fc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.700282] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.957423] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319144, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.009927] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319146, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.051829] env[62521]: DEBUG oslo_vmware.api [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319147, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218676} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.052163] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1066.052311] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1066.053069] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1066.053069] env[62521]: INFO nova.compute.manager [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1066.053069] env[62521]: DEBUG oslo.service.loopingcall [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.053198] env[62521]: DEBUG nova.compute.manager [-] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1066.053198] env[62521]: DEBUG nova.network.neutron [-] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1066.101434] env[62521]: INFO nova.compute.manager [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Took 12.72 seconds to build instance. [ 1066.114796] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5264cd02-ab4e-c3d3-f1aa-76436c5bb8a4, 'name': SearchDatastore_Task, 'duration_secs': 0.009786} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.115134] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.115379] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1066.115871] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.116218] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.116330] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1066.116594] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a92ddf6-e165-4341-9f97-f09025c1fb60 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.124993] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1066.125225] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1066.126538] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c1751c3-613c-43ad-ab25-7f07467e8e05 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.132478] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1066.132478] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f877ac-6e29-1bae-e9d1-5ded7dc30bb8" [ 1066.132478] env[62521]: _type = "Task" [ 1066.132478] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.144042] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f877ac-6e29-1bae-e9d1-5ded7dc30bb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.205649] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1066.205987] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.288s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.206155] env[62521]: DEBUG oslo_concurrency.lockutils [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.597s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.206417] env[62521]: DEBUG nova.objects.instance [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lazy-loading 'resources' on Instance uuid 8b4db43f-bc21-4d40-9410-f643971e53fc {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.212317] env[62521]: DEBUG nova.network.neutron [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Updated VIF entry in instance network info cache for port f4903211-c60f-4fa2-9589-a98babc4b5c5. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1066.212882] env[62521]: DEBUG nova.network.neutron [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Updating instance_info_cache with network_info: [{"id": "f4903211-c60f-4fa2-9589-a98babc4b5c5", "address": "fa:16:3e:a2:d7:bd", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4903211-c6", "ovs_interfaceid": "f4903211-c60f-4fa2-9589-a98babc4b5c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.230509] env[62521]: DEBUG oslo_concurrency.lockutils [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "bc7d76b1-76b2-49a7-8188-0672312767a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.230815] env[62521]: DEBUG oslo_concurrency.lockutils [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "bc7d76b1-76b2-49a7-8188-0672312767a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.231043] env[62521]: DEBUG oslo_concurrency.lockutils [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "bc7d76b1-76b2-49a7-8188-0672312767a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.231243] env[62521]: DEBUG oslo_concurrency.lockutils [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "bc7d76b1-76b2-49a7-8188-0672312767a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.231418] env[62521]: DEBUG oslo_concurrency.lockutils [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "bc7d76b1-76b2-49a7-8188-0672312767a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.233984] env[62521]: INFO nova.compute.manager [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Terminating instance [ 1066.236199] env[62521]: DEBUG nova.compute.manager [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1066.236424] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1066.237397] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cbc3670-a7a6-4bc6-99c7-d90b58fbc9b0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.246509] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1066.246792] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9981375c-c83e-48b9-812b-25e2198d92c0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.311067] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1066.311266] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1066.311557] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleting the datastore file [datastore2] bc7d76b1-76b2-49a7-8188-0672312767a1 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1066.311742] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75b46614-abab-4c78-ac4c-979dc9716916 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.318323] env[62521]: DEBUG oslo_vmware.api [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1066.318323] env[62521]: value = "task-1319149" [ 1066.318323] env[62521]: _type = "Task" [ 1066.318323] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.325982] env[62521]: DEBUG oslo_vmware.api [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319149, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.436776] env[62521]: DEBUG nova.network.neutron [req-1f86e073-d523-4f84-99a5-1d3357649f4d req-6394b70b-2bbf-4cf0-a06e-e82891d21127 service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Updated VIF entry in instance network info cache for port 9ee94c09-d3e5-465a-9926-639300512161. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1066.437334] env[62521]: DEBUG nova.network.neutron [req-1f86e073-d523-4f84-99a5-1d3357649f4d req-6394b70b-2bbf-4cf0-a06e-e82891d21127 service nova] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Updating instance_info_cache with network_info: [{"id": "9ee94c09-d3e5-465a-9926-639300512161", "address": "fa:16:3e:0b:5a:3b", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": null, "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9ee94c09-d3", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.457083] env[62521]: DEBUG oslo_vmware.api [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319144, 'name': PowerOnVM_Task, 'duration_secs': 0.616868} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.457354] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1066.459889] env[62521]: DEBUG nova.compute.manager [None req-4d768430-0e42-4bde-900c-e6873eb95788 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1066.460674] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a3ba100-4301-405b-9c19-12d888ecc074 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.509732] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319146, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.603656] env[62521]: DEBUG oslo_concurrency.lockutils [None req-67da8e9b-7372-4e1b-b2e2-9253c59e70ae tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.229s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.645727] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f877ac-6e29-1bae-e9d1-5ded7dc30bb8, 'name': SearchDatastore_Task, 'duration_secs': 0.011797} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.646536] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97ff46c9-a94a-4212-bbe6-c1cd4763419b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.652263] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1066.652263] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5258e132-596a-b94a-3040-ee2477637c5b" [ 1066.652263] env[62521]: _type = "Task" [ 1066.652263] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.660794] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5258e132-596a-b94a-3040-ee2477637c5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.709683] env[62521]: DEBUG nova.objects.instance [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lazy-loading 'numa_topology' on Instance uuid 8b4db43f-bc21-4d40-9410-f643971e53fc {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.715935] env[62521]: DEBUG oslo_concurrency.lockutils [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] Releasing lock "refresh_cache-3aa60da8-b251-4f31-bf8d-a4cd143b491e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.716133] env[62521]: DEBUG nova.compute.manager [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received event network-changed-2beae869-6010-48ba-8e25-6b392dd2ec1f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.716323] env[62521]: DEBUG nova.compute.manager [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Refreshing instance network info cache due to event network-changed-2beae869-6010-48ba-8e25-6b392dd2ec1f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1066.716549] env[62521]: DEBUG oslo_concurrency.lockutils [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] Acquiring lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.716696] env[62521]: DEBUG oslo_concurrency.lockutils [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] Acquired lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.716860] env[62521]: DEBUG nova.network.neutron [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Refreshing network info cache for port 2beae869-6010-48ba-8e25-6b392dd2ec1f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1066.828637] env[62521]: DEBUG oslo_vmware.api [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319149, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.940932] env[62521]: DEBUG oslo_concurrency.lockutils [req-1f86e073-d523-4f84-99a5-1d3357649f4d req-6394b70b-2bbf-4cf0-a06e-e82891d21127 service nova] Releasing lock "refresh_cache-8b4db43f-bc21-4d40-9410-f643971e53fc" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.013118] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319146, 'name': CloneVM_Task} progress is 95%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.170239] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5258e132-596a-b94a-3040-ee2477637c5b, 'name': SearchDatastore_Task, 'duration_secs': 0.013239} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.170239] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.170360] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 3aa60da8-b251-4f31-bf8d-a4cd143b491e/3aa60da8-b251-4f31-bf8d-a4cd143b491e.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1067.170805] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bea9fa09-f0c1-4035-9300-204914674cd8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.178330] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1067.178330] env[62521]: value = "task-1319150" [ 1067.178330] env[62521]: _type = "Task" [ 1067.178330] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.186344] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319150, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.212098] env[62521]: DEBUG nova.objects.base [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Object Instance<8b4db43f-bc21-4d40-9410-f643971e53fc> lazy-loaded attributes: resources,numa_topology {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1067.334380] env[62521]: DEBUG oslo_vmware.api [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319149, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.566602} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.336988] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1067.337203] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1067.337390] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1067.337574] env[62521]: INFO nova.compute.manager [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1067.337822] env[62521]: DEBUG oslo.service.loopingcall [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.339109] env[62521]: DEBUG nova.compute.manager [-] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1067.339109] env[62521]: DEBUG nova.network.neutron [-] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1067.360129] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570916fd-c667-4636-a2bd-e442dbc2b59a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.367862] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7eef3c1-f469-4a4c-bbfc-a4d2b603c721 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.401526] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889daa03-91ae-475c-932f-e87184fd4827 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.410107] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd695045-fabe-443b-82d6-99d6eaf77f9b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.428028] env[62521]: DEBUG nova.compute.provider_tree [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.513767] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319146, 'name': CloneVM_Task, 'duration_secs': 1.695534} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.514256] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Created linked-clone VM from snapshot [ 1067.515269] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8ce9b4-322a-453b-99b3-609f01b93060 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.518374] env[62521]: DEBUG nova.network.neutron [-] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.527110] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Uploading image 2adddf14-344e-4adb-9e63-62c5a1ab8f27 {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1067.536793] env[62521]: DEBUG nova.compute.manager [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Received event network-changed-c6383189-f73f-4f87-8424-e8956b34f94c {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.537759] env[62521]: DEBUG nova.compute.manager [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Refreshing instance network info cache due to event network-changed-c6383189-f73f-4f87-8424-e8956b34f94c. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1067.537989] env[62521]: DEBUG oslo_concurrency.lockutils [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] Acquiring lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.538180] env[62521]: DEBUG oslo_concurrency.lockutils [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] Acquired lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.538407] env[62521]: DEBUG nova.network.neutron [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Refreshing network info cache for port c6383189-f73f-4f87-8424-e8956b34f94c {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.546606] env[62521]: DEBUG nova.network.neutron [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updated VIF entry in instance network info cache for port 2beae869-6010-48ba-8e25-6b392dd2ec1f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1067.546606] env[62521]: DEBUG nova.network.neutron [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.572681] env[62521]: DEBUG oslo_vmware.rw_handles [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1067.572681] env[62521]: value = "vm-282190" [ 1067.572681] env[62521]: _type = "VirtualMachine" [ 1067.572681] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1067.573856] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c02e320d-8016-4432-a08f-15f2c8fcd809 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.584206] env[62521]: DEBUG oslo_vmware.rw_handles [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lease: (returnval){ [ 1067.584206] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e5fd7-57ce-b683-f73d-2170020d9eae" [ 1067.584206] env[62521]: _type = "HttpNfcLease" [ 1067.584206] env[62521]: } obtained for exporting VM: (result){ [ 1067.584206] env[62521]: value = "vm-282190" [ 1067.584206] env[62521]: _type = "VirtualMachine" [ 1067.584206] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1067.584206] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the lease: (returnval){ [ 1067.584206] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e5fd7-57ce-b683-f73d-2170020d9eae" [ 1067.584206] env[62521]: _type = "HttpNfcLease" [ 1067.584206] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1067.592182] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1067.592182] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e5fd7-57ce-b683-f73d-2170020d9eae" [ 1067.592182] env[62521]: _type = "HttpNfcLease" [ 1067.592182] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1067.689268] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319150, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.807527] env[62521]: DEBUG nova.compute.manager [req-2f3b81d4-bd80-419f-ab8b-9ffb6cb49419 req-f9f04b8b-8e06-4a37-afbf-bb8afebaab3b service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Received event network-vif-deleted-e02e66bc-8bdd-4eed-81a7-612a3d86e2f7 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.807831] env[62521]: INFO nova.compute.manager [req-2f3b81d4-bd80-419f-ab8b-9ffb6cb49419 req-f9f04b8b-8e06-4a37-afbf-bb8afebaab3b service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Neutron deleted interface e02e66bc-8bdd-4eed-81a7-612a3d86e2f7; detaching it from the instance and deleting it from the info cache [ 1067.807979] env[62521]: DEBUG nova.network.neutron [req-2f3b81d4-bd80-419f-ab8b-9ffb6cb49419 req-f9f04b8b-8e06-4a37-afbf-bb8afebaab3b service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.931530] env[62521]: DEBUG nova.scheduler.client.report [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.023618] env[62521]: INFO nova.compute.manager [-] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Took 1.97 seconds to deallocate network for instance. [ 1068.049776] env[62521]: DEBUG oslo_concurrency.lockutils [req-a583b932-67fe-40b6-92de-1c6560b0f89d req-771caf13-c2ce-40bf-9fb4-e378b16a934e service nova] Releasing lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.095570] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1068.095570] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e5fd7-57ce-b683-f73d-2170020d9eae" [ 1068.095570] env[62521]: _type = "HttpNfcLease" [ 1068.095570] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1068.095999] env[62521]: DEBUG oslo_vmware.rw_handles [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1068.095999] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524e5fd7-57ce-b683-f73d-2170020d9eae" [ 1068.095999] env[62521]: _type = "HttpNfcLease" [ 1068.095999] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1068.096855] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e768d306-982b-4836-bc46-8348a8191b93 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.104659] env[62521]: DEBUG oslo_vmware.rw_handles [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2c6b8-2f4a-a106-a03c-3155b5b15f96/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1068.104850] env[62521]: DEBUG oslo_vmware.rw_handles [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2c6b8-2f4a-a106-a03c-3155b5b15f96/disk-0.vmdk for reading. {{(pid=62521) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1068.196092] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319150, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.61852} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.196092] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 3aa60da8-b251-4f31-bf8d-a4cd143b491e/3aa60da8-b251-4f31-bf8d-a4cd143b491e.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1068.196092] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1068.196092] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-995ab164-d1dc-4432-8a49-dc70992942de {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.204500] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1068.204500] env[62521]: value = "task-1319152" [ 1068.204500] env[62521]: _type = "Task" [ 1068.204500] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.211744] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319152, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.216103] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e97c49a8-d529-4033-8696-6fb2b92a1bc3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.279571] env[62521]: DEBUG nova.network.neutron [-] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.312714] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-544dfaba-37e9-4f7f-9254-528388aa6918 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.324339] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b08fb7a-0161-4b64-b01f-bdda1ad82129 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.352837] env[62521]: DEBUG nova.compute.manager [req-2f3b81d4-bd80-419f-ab8b-9ffb6cb49419 req-f9f04b8b-8e06-4a37-afbf-bb8afebaab3b service nova] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Detach interface failed, port_id=e02e66bc-8bdd-4eed-81a7-612a3d86e2f7, reason: Instance bc7d76b1-76b2-49a7-8188-0672312767a1 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1068.422340] env[62521]: DEBUG nova.network.neutron [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updated VIF entry in instance network info cache for port c6383189-f73f-4f87-8424-e8956b34f94c. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1068.422724] env[62521]: DEBUG nova.network.neutron [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance_info_cache with network_info: [{"id": "c6383189-f73f-4f87-8424-e8956b34f94c", "address": "fa:16:3e:ba:1a:11", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6383189-f7", "ovs_interfaceid": "c6383189-f73f-4f87-8424-e8956b34f94c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.438048] env[62521]: DEBUG oslo_concurrency.lockutils [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.232s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.530246] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.530538] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.530771] env[62521]: DEBUG nova.objects.instance [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'resources' on Instance uuid 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.712459] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319152, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091956} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.712459] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.713195] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3bbc52-71a2-48f4-a8fd-ed81b274d7e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.735442] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 3aa60da8-b251-4f31-bf8d-a4cd143b491e/3aa60da8-b251-4f31-bf8d-a4cd143b491e.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.735738] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7c7b2ac-9e58-496c-9924-a4ae2d0a16f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.755736] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1068.755736] env[62521]: value = "task-1319153" [ 1068.755736] env[62521]: _type = "Task" [ 1068.755736] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.764252] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319153, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.781945] env[62521]: INFO nova.compute.manager [-] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Took 1.44 seconds to deallocate network for instance. [ 1068.925071] env[62521]: DEBUG oslo_concurrency.lockutils [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] Releasing lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.925406] env[62521]: DEBUG nova.compute.manager [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Received event network-vif-deleted-8824a231-01df-41e5-9bea-9071d8b41cb8 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.925601] env[62521]: INFO nova.compute.manager [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Neutron deleted interface 8824a231-01df-41e5-9bea-9071d8b41cb8; detaching it from the instance and deleting it from the info cache [ 1068.925799] env[62521]: DEBUG nova.network.neutron [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.948399] env[62521]: DEBUG oslo_concurrency.lockutils [None req-865f9a93-a734-45b1-a1a3-abebf36ec7c3 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 25.608s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.949734] env[62521]: DEBUG oslo_concurrency.lockutils [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 3.262s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.949943] env[62521]: DEBUG oslo_concurrency.lockutils [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "8b4db43f-bc21-4d40-9410-f643971e53fc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.950560] env[62521]: DEBUG oslo_concurrency.lockutils [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.950560] env[62521]: DEBUG oslo_concurrency.lockutils [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.953863] env[62521]: INFO nova.compute.manager [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Terminating instance [ 1068.958908] env[62521]: DEBUG nova.compute.manager [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1068.958908] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1068.959057] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f3d084a6-f788-447a-ac01-6c57971627d5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.968991] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1423c895-3518-4aee-b725-5c1d2a817d71 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.005823] env[62521]: WARNING nova.virt.vmwareapi.vmops [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8b4db43f-bc21-4d40-9410-f643971e53fc could not be found. [ 1069.006149] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.006723] env[62521]: INFO nova.compute.manager [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1069.006905] env[62521]: DEBUG oslo.service.loopingcall [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.007479] env[62521]: DEBUG nova.compute.manager [-] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1069.007862] env[62521]: DEBUG nova.network.neutron [-] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1069.186200] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878d20cf-7d87-4215-9d84-fe762a7d4e1d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.194822] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989fd636-4791-47b2-802b-a81b8e5794ac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.226108] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09fefc65-6e6e-4887-a8b9-7d0310b1a73f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.234609] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990adc9f-a2fc-493a-a123-e2fdf9c79c24 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.249729] env[62521]: DEBUG nova.compute.provider_tree [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.267888] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319153, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.288639] env[62521]: DEBUG oslo_concurrency.lockutils [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.432018] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5b323f56-9263-46af-83a2-5a184be75431 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.440126] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6fd5cc-9eec-4c79-bf88-5eeebf00f379 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.469014] env[62521]: DEBUG nova.compute.manager [req-85f3d797-1550-48ca-ad90-be2833647313 req-801a95ba-37d7-4a8c-b28a-f533e1e41ce4 service nova] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Detach interface failed, port_id=8824a231-01df-41e5-9bea-9071d8b41cb8, reason: Instance 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1069.487947] env[62521]: INFO nova.compute.manager [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Rescuing [ 1069.488308] env[62521]: DEBUG oslo_concurrency.lockutils [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.488609] env[62521]: DEBUG oslo_concurrency.lockutils [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.488793] env[62521]: DEBUG nova.network.neutron [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1069.754208] env[62521]: DEBUG nova.scheduler.client.report [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1069.767196] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319153, 'name': ReconfigVM_Task, 'duration_secs': 0.637403} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.767562] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 3aa60da8-b251-4f31-bf8d-a4cd143b491e/3aa60da8-b251-4f31-bf8d-a4cd143b491e.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.768234] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-31cdf4bc-e33c-42cc-bbb7-dad7f3e55697 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.775286] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1069.775286] env[62521]: value = "task-1319154" [ 1069.775286] env[62521]: _type = "Task" [ 1069.775286] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.783491] env[62521]: DEBUG nova.network.neutron [-] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.784712] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319154, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.227856] env[62521]: DEBUG nova.network.neutron [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Updating instance_info_cache with network_info: [{"id": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "address": "fa:16:3e:44:95:f8", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee2404e8-7a", "ovs_interfaceid": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.260241] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.262748] env[62521]: DEBUG oslo_concurrency.lockutils [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.974s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.263008] env[62521]: DEBUG nova.objects.instance [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lazy-loading 'resources' on Instance uuid bc7d76b1-76b2-49a7-8188-0672312767a1 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.285558] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319154, 'name': Rename_Task, 'duration_secs': 0.253626} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.285901] env[62521]: INFO nova.compute.manager [-] [instance: 8b4db43f-bc21-4d40-9410-f643971e53fc] Took 1.28 seconds to deallocate network for instance. [ 1070.286371] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1070.287889] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbbe3bf4-0223-4efd-bdb8-780e3aaff096 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.293626] env[62521]: INFO nova.scheduler.client.report [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleted allocations for instance 7b5afeab-f04e-453a-a3ed-09e3a725e8d6 [ 1070.304452] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1070.304452] env[62521]: value = "task-1319155" [ 1070.304452] env[62521]: _type = "Task" [ 1070.304452] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.313404] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319155, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.730712] env[62521]: DEBUG oslo_concurrency.lockutils [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.803453] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8153d826-919a-4760-9114-2ba920a68f2e tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "7b5afeab-f04e-453a-a3ed-09e3a725e8d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.911s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.822687] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319155, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.888443] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb3be00-451d-4b16-85c9-64e23662ad38 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.897452] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ebe61a-9b19-402f-972c-0af69f2d315d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.932068] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5859c161-62e6-4821-8e4a-d211bb653c41 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.941411] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab97718-59e7-4e51-b878-38dfd6d9e807 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.955633] env[62521]: DEBUG nova.compute.provider_tree [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.263611] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1071.263611] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8be287cd-f53b-439e-90eb-ee7b8303a0e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.271034] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1071.271034] env[62521]: value = "task-1319156" [ 1071.271034] env[62521]: _type = "Task" [ 1071.271034] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.279580] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319156, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.315453] env[62521]: DEBUG oslo_vmware.api [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319155, 'name': PowerOnVM_Task, 'duration_secs': 0.606369} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.315745] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1071.315959] env[62521]: INFO nova.compute.manager [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Took 8.45 seconds to spawn the instance on the hypervisor. [ 1071.316159] env[62521]: DEBUG nova.compute.manager [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1071.316970] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d4336f-d0e0-488f-a2f3-f2726cd8ad83 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.322855] env[62521]: DEBUG oslo_concurrency.lockutils [None req-19fda9a4-55cf-41c5-a3ea-e57491f769f7 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "8b4db43f-bc21-4d40-9410-f643971e53fc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.373s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.459605] env[62521]: DEBUG nova.scheduler.client.report [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.780669] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319156, 'name': PowerOffVM_Task, 'duration_secs': 0.421518} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.781059] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.781841] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9f0235-4038-4b33-a003-fb3a1ebe8e92 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.800103] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd9ec5d-d490-455f-888d-db80d1dbd5e0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.830443] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1071.830985] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2583b37c-2e27-4614-9816-df4a21a6dc13 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.841127] env[62521]: INFO nova.compute.manager [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Took 14.86 seconds to build instance. [ 1071.843147] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1071.843147] env[62521]: value = "task-1319157" [ 1071.843147] env[62521]: _type = "Task" [ 1071.843147] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.852626] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1071.852891] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1071.853156] env[62521]: DEBUG oslo_concurrency.lockutils [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.853311] env[62521]: DEBUG oslo_concurrency.lockutils [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.853495] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1071.853756] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94fdd24a-e95d-498d-b5ac-e1c7d7c27afd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.863753] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1071.863963] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1071.864721] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1daf075e-70d8-4551-9623-18527f8f3177 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.870326] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1071.870326] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5292bd06-3c1d-5e15-9208-ba4e32e07189" [ 1071.870326] env[62521]: _type = "Task" [ 1071.870326] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.878463] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5292bd06-3c1d-5e15-9208-ba4e32e07189, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.964956] env[62521]: DEBUG oslo_concurrency.lockutils [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.702s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.989184] env[62521]: INFO nova.scheduler.client.report [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted allocations for instance bc7d76b1-76b2-49a7-8188-0672312767a1 [ 1072.067918] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "fbb2c755-3289-453d-ad19-40cc53b56fa2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.068433] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "fbb2c755-3289-453d-ad19-40cc53b56fa2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.132846] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "e159807b-c7b0-4d2e-a7be-426ecaf01785" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.133166] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.344187] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3e6e7c4a-c466-4640-acb6-ccef74c170f1 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.375s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.381059] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5292bd06-3c1d-5e15-9208-ba4e32e07189, 'name': SearchDatastore_Task, 'duration_secs': 0.031239} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.381850] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b02ea0c-4e09-4191-a73d-8aa49dce637e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.386812] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1072.386812] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525f52e4-3473-97cc-6a2e-c389166293e0" [ 1072.386812] env[62521]: _type = "Task" [ 1072.386812] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.396250] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525f52e4-3473-97cc-6a2e-c389166293e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.493894] env[62521]: DEBUG nova.compute.manager [req-5a402d20-004e-4333-9e61-4c9d0fa050b4 req-b9b95320-7659-4603-8b8f-a385231ad909 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Received event network-changed-f4903211-c60f-4fa2-9589-a98babc4b5c5 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.494135] env[62521]: DEBUG nova.compute.manager [req-5a402d20-004e-4333-9e61-4c9d0fa050b4 req-b9b95320-7659-4603-8b8f-a385231ad909 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Refreshing instance network info cache due to event network-changed-f4903211-c60f-4fa2-9589-a98babc4b5c5. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1072.494363] env[62521]: DEBUG oslo_concurrency.lockutils [req-5a402d20-004e-4333-9e61-4c9d0fa050b4 req-b9b95320-7659-4603-8b8f-a385231ad909 service nova] Acquiring lock "refresh_cache-3aa60da8-b251-4f31-bf8d-a4cd143b491e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.494515] env[62521]: DEBUG oslo_concurrency.lockutils [req-5a402d20-004e-4333-9e61-4c9d0fa050b4 req-b9b95320-7659-4603-8b8f-a385231ad909 service nova] Acquired lock "refresh_cache-3aa60da8-b251-4f31-bf8d-a4cd143b491e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.494686] env[62521]: DEBUG nova.network.neutron [req-5a402d20-004e-4333-9e61-4c9d0fa050b4 req-b9b95320-7659-4603-8b8f-a385231ad909 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Refreshing network info cache for port f4903211-c60f-4fa2-9589-a98babc4b5c5 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1072.500049] env[62521]: DEBUG oslo_concurrency.lockutils [None req-79a243b5-277f-455c-979e-f17f0ad91863 tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "bc7d76b1-76b2-49a7-8188-0672312767a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.269s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.571683] env[62521]: DEBUG nova.compute.manager [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1072.635576] env[62521]: DEBUG nova.compute.manager [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1072.899014] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525f52e4-3473-97cc-6a2e-c389166293e0, 'name': SearchDatastore_Task, 'duration_secs': 0.035309} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.899374] env[62521]: DEBUG oslo_concurrency.lockutils [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.899713] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 703ff423-98b1-4942-b6fd-8b95fe57bd0a/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk. {{(pid=62521) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1072.900030] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2de275df-f66d-4d0c-a462-50268302075d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.908341] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1072.908341] env[62521]: value = "task-1319158" [ 1072.908341] env[62521]: _type = "Task" [ 1072.908341] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.916992] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319158, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.098371] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.098729] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.100283] env[62521]: INFO nova.compute.claims [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1073.160090] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.276071] env[62521]: DEBUG nova.network.neutron [req-5a402d20-004e-4333-9e61-4c9d0fa050b4 req-b9b95320-7659-4603-8b8f-a385231ad909 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Updated VIF entry in instance network info cache for port f4903211-c60f-4fa2-9589-a98babc4b5c5. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1073.276541] env[62521]: DEBUG nova.network.neutron [req-5a402d20-004e-4333-9e61-4c9d0fa050b4 req-b9b95320-7659-4603-8b8f-a385231ad909 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Updating instance_info_cache with network_info: [{"id": "f4903211-c60f-4fa2-9589-a98babc4b5c5", "address": "fa:16:3e:a2:d7:bd", "network": {"id": "9676b0ee-cfd5-4453-86dd-ff916038f0ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-709324203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5780fc8ecfd64161bc2ea0b3a092b4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4903211-c6", "ovs_interfaceid": "f4903211-c60f-4fa2-9589-a98babc4b5c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.420021] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319158, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.780487] env[62521]: DEBUG oslo_concurrency.lockutils [req-5a402d20-004e-4333-9e61-4c9d0fa050b4 req-b9b95320-7659-4603-8b8f-a385231ad909 service nova] Releasing lock "refresh_cache-3aa60da8-b251-4f31-bf8d-a4cd143b491e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.919348] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319158, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.652392} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.919542] env[62521]: INFO nova.virt.vmwareapi.ds_util [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 703ff423-98b1-4942-b6fd-8b95fe57bd0a/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk. [ 1073.920200] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a08a40-cc39-4a38-bf16-2b0b95254b0a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.945977] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 703ff423-98b1-4942-b6fd-8b95fe57bd0a/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.946331] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a94f9b4f-3393-4577-895e-ec8a99c82fad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.959124] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "d574f077-b221-4f91-8b54-0915421cb36f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.959371] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "d574f077-b221-4f91-8b54-0915421cb36f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.959579] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "d574f077-b221-4f91-8b54-0915421cb36f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.959818] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "d574f077-b221-4f91-8b54-0915421cb36f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.959934] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "d574f077-b221-4f91-8b54-0915421cb36f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.962477] env[62521]: INFO nova.compute.manager [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Terminating instance [ 1073.964314] env[62521]: DEBUG nova.compute.manager [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1073.964512] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1073.965584] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b0bb60-e5d2-4853-96b8-f181348fc097 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.969467] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1073.969467] env[62521]: value = "task-1319159" [ 1073.969467] env[62521]: _type = "Task" [ 1073.969467] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.974983] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1073.975565] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f69d625-7483-4510-b03f-443ea8506797 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.980152] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319159, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.984127] env[62521]: DEBUG oslo_vmware.api [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1073.984127] env[62521]: value = "task-1319160" [ 1073.984127] env[62521]: _type = "Task" [ 1073.984127] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.992716] env[62521]: DEBUG oslo_vmware.api [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319160, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.250975] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a0e5b6-7aa2-49e7-965d-d212e84e0d40 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.258940] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ea6f45-9bc3-4221-88b9-db8f1d8215f0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.288746] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cc0c71-860e-4e92-95d5-dc3ef2b6d31f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.296752] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51110fa0-f5e5-40f9-b69e-0fae983340ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.310440] env[62521]: DEBUG nova.compute.provider_tree [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.481386] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319159, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.493056] env[62521]: DEBUG oslo_vmware.api [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319160, 'name': PowerOffVM_Task, 'duration_secs': 0.203647} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.493356] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1074.493532] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1074.493796] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9b399a7-8330-4aac-97a8-e650652e3346 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.556025] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1074.556400] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1074.556679] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleting the datastore file [datastore2] d574f077-b221-4f91-8b54-0915421cb36f {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.557062] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fcfcd089-6ad4-4d3e-9b6f-da90eb19b959 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.563577] env[62521]: DEBUG oslo_vmware.api [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for the task: (returnval){ [ 1074.563577] env[62521]: value = "task-1319162" [ 1074.563577] env[62521]: _type = "Task" [ 1074.563577] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.571179] env[62521]: DEBUG oslo_vmware.api [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319162, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.814395] env[62521]: DEBUG nova.scheduler.client.report [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.980238] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319159, 'name': ReconfigVM_Task, 'duration_secs': 0.518378} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.980482] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 703ff423-98b1-4942-b6fd-8b95fe57bd0a/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.981345] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8588336b-2b96-4197-89ea-9eed7bc474fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.009345] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95292695-0018-47e2-b50f-6bba37fc5446 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.025787] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1075.025787] env[62521]: value = "task-1319163" [ 1075.025787] env[62521]: _type = "Task" [ 1075.025787] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.034240] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319163, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.074605] env[62521]: DEBUG oslo_vmware.api [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Task: {'id': task-1319162, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.312027} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.074903] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.075112] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1075.075299] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1075.075479] env[62521]: INFO nova.compute.manager [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1075.075731] env[62521]: DEBUG oslo.service.loopingcall [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.075929] env[62521]: DEBUG nova.compute.manager [-] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1075.076034] env[62521]: DEBUG nova.network.neutron [-] [instance: d574f077-b221-4f91-8b54-0915421cb36f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1075.321190] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.221s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.321190] env[62521]: DEBUG nova.compute.manager [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1075.323966] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.164s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.325974] env[62521]: INFO nova.compute.claims [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1075.384005] env[62521]: DEBUG nova.compute.manager [req-904c63b4-9747-4615-bbef-74dcf5b224a2 req-86ca971a-a696-4693-9910-0e1be07b0b5d service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Received event network-vif-deleted-862343de-1cfa-4880-832a-bc8881cd0c30 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.384241] env[62521]: INFO nova.compute.manager [req-904c63b4-9747-4615-bbef-74dcf5b224a2 req-86ca971a-a696-4693-9910-0e1be07b0b5d service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Neutron deleted interface 862343de-1cfa-4880-832a-bc8881cd0c30; detaching it from the instance and deleting it from the info cache [ 1075.384415] env[62521]: DEBUG nova.network.neutron [req-904c63b4-9747-4615-bbef-74dcf5b224a2 req-86ca971a-a696-4693-9910-0e1be07b0b5d service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.536249] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319163, 'name': ReconfigVM_Task, 'duration_secs': 0.323556} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.536611] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1075.536790] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07e929e1-fd2f-4eed-a591-52c488222988 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.544029] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1075.544029] env[62521]: value = "task-1319164" [ 1075.544029] env[62521]: _type = "Task" [ 1075.544029] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.551897] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319164, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.830226] env[62521]: DEBUG nova.compute.utils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1075.831742] env[62521]: DEBUG nova.compute.manager [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1075.831944] env[62521]: DEBUG nova.network.neutron [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1075.845142] env[62521]: DEBUG nova.network.neutron [-] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.875812] env[62521]: DEBUG nova.policy [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3566601952144168b5d400c41026c81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0dfb5f5bd4646ec95b7c6a2d0434175', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1075.887175] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c40a645-974c-4c6b-bde1-34f7c3dbe19b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.901786] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242ab571-0d63-40be-940c-3e92036a3ebc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.944251] env[62521]: DEBUG nova.compute.manager [req-904c63b4-9747-4615-bbef-74dcf5b224a2 req-86ca971a-a696-4693-9910-0e1be07b0b5d service nova] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Detach interface failed, port_id=862343de-1cfa-4880-832a-bc8881cd0c30, reason: Instance d574f077-b221-4f91-8b54-0915421cb36f could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1076.055669] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319164, 'name': PowerOnVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.264232] env[62521]: DEBUG nova.network.neutron [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Successfully created port: 51638ad7-153a-49c6-8d83-8efc154d8502 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1076.337677] env[62521]: DEBUG nova.compute.manager [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1076.364034] env[62521]: INFO nova.compute.manager [-] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Took 1.29 seconds to deallocate network for instance. [ 1076.505353] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a7b48f-dfce-464d-9ed0-d153e90719f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.515637] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26235044-b152-4dd7-9c4a-09c3be5566e9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.561804] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85346ce8-a71d-46eb-ae86-b7ded3488941 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.569490] env[62521]: DEBUG oslo_vmware.api [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319164, 'name': PowerOnVM_Task, 'duration_secs': 0.855181} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.571451] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1076.574106] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db37fed1-1bd0-442b-bac6-dfdf9885ecce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.578633] env[62521]: DEBUG nova.compute.manager [None req-315825a9-34d0-4d30-8db2-0c9aa0c05ee7 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1076.579413] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba53ff9-12a7-4a91-8118-f8c712aa6b23 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.595695] env[62521]: DEBUG nova.compute.provider_tree [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.875682] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.101137] env[62521]: DEBUG nova.scheduler.client.report [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1077.344837] env[62521]: DEBUG nova.compute.manager [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1077.370772] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.371082] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.371258] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.371451] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.371604] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.372358] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.372605] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.372778] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.372952] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.373139] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.373325] env[62521]: DEBUG nova.virt.hardware [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.374509] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba7632c-dd3a-47d0-b2ee-15618a55a47e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.384273] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-441b9840-2095-47bd-a409-4bc1b63b5322 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.607164] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.607687] env[62521]: DEBUG nova.compute.manager [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1077.610503] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.736s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.610830] env[62521]: DEBUG nova.objects.instance [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lazy-loading 'resources' on Instance uuid d574f077-b221-4f91-8b54-0915421cb36f {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.685285] env[62521]: DEBUG nova.compute.manager [req-d77623bf-b523-49a4-94ae-7f1572d8edff req-a366921c-3733-4332-8e88-b5ffad7c9450 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Received event network-vif-plugged-51638ad7-153a-49c6-8d83-8efc154d8502 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1077.686017] env[62521]: DEBUG oslo_concurrency.lockutils [req-d77623bf-b523-49a4-94ae-7f1572d8edff req-a366921c-3733-4332-8e88-b5ffad7c9450 service nova] Acquiring lock "fbb2c755-3289-453d-ad19-40cc53b56fa2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.686017] env[62521]: DEBUG oslo_concurrency.lockutils [req-d77623bf-b523-49a4-94ae-7f1572d8edff req-a366921c-3733-4332-8e88-b5ffad7c9450 service nova] Lock "fbb2c755-3289-453d-ad19-40cc53b56fa2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.687461] env[62521]: DEBUG oslo_concurrency.lockutils [req-d77623bf-b523-49a4-94ae-7f1572d8edff req-a366921c-3733-4332-8e88-b5ffad7c9450 service nova] Lock "fbb2c755-3289-453d-ad19-40cc53b56fa2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.687461] env[62521]: DEBUG nova.compute.manager [req-d77623bf-b523-49a4-94ae-7f1572d8edff req-a366921c-3733-4332-8e88-b5ffad7c9450 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] No waiting events found dispatching network-vif-plugged-51638ad7-153a-49c6-8d83-8efc154d8502 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1077.687461] env[62521]: WARNING nova.compute.manager [req-d77623bf-b523-49a4-94ae-7f1572d8edff req-a366921c-3733-4332-8e88-b5ffad7c9450 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Received unexpected event network-vif-plugged-51638ad7-153a-49c6-8d83-8efc154d8502 for instance with vm_state building and task_state spawning. [ 1077.781036] env[62521]: DEBUG nova.network.neutron [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Successfully updated port: 51638ad7-153a-49c6-8d83-8efc154d8502 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1077.837882] env[62521]: INFO nova.compute.manager [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Unrescuing [ 1077.838179] env[62521]: DEBUG oslo_concurrency.lockutils [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.838338] env[62521]: DEBUG oslo_concurrency.lockutils [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.838544] env[62521]: DEBUG nova.network.neutron [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1077.845291] env[62521]: DEBUG oslo_vmware.rw_handles [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2c6b8-2f4a-a106-a03c-3155b5b15f96/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1077.846245] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9610f322-aa36-4d72-9adf-825b8c776594 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.853421] env[62521]: DEBUG oslo_vmware.rw_handles [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2c6b8-2f4a-a106-a03c-3155b5b15f96/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1077.853594] env[62521]: ERROR oslo_vmware.rw_handles [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2c6b8-2f4a-a106-a03c-3155b5b15f96/disk-0.vmdk due to incomplete transfer. [ 1077.853822] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f3d6ac2f-65eb-430d-b55d-5dad08086f38 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.860963] env[62521]: DEBUG oslo_vmware.rw_handles [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2c6b8-2f4a-a106-a03c-3155b5b15f96/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1077.861187] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Uploaded image 2adddf14-344e-4adb-9e63-62c5a1ab8f27 to the Glance image server {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1077.863457] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1077.863703] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d60ee862-2242-4a88-bd39-53b75aadee61 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.869298] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1077.869298] env[62521]: value = "task-1319165" [ 1077.869298] env[62521]: _type = "Task" [ 1077.869298] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.876995] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319165, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.115135] env[62521]: DEBUG nova.compute.utils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1078.119532] env[62521]: DEBUG nova.compute.manager [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1078.119884] env[62521]: DEBUG nova.network.neutron [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1078.160494] env[62521]: DEBUG nova.policy [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a0e35b0e4574ed9afc4549d30daa210', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '31c788848b0d47478564e53066e7c51a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1078.242975] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b1059b-77bc-4f34-94a0-7b6fafd35c4b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.251189] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c2cc2a-c73e-46b1-a6eb-30b491ab020e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.280999] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc698a4-d02d-4869-96a0-b6524040197c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.287777] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "refresh_cache-fbb2c755-3289-453d-ad19-40cc53b56fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.287916] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "refresh_cache-fbb2c755-3289-453d-ad19-40cc53b56fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.288080] env[62521]: DEBUG nova.network.neutron [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1078.290148] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb9c268-9853-4457-aa9f-fa48c8f291aa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.305424] env[62521]: DEBUG nova.compute.provider_tree [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.378821] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319165, 'name': Destroy_Task, 'duration_secs': 0.345837} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.379043] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Destroyed the VM [ 1078.379284] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1078.379827] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c12bde4a-04f0-4ffe-b835-45cc7e213f45 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.385866] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1078.385866] env[62521]: value = "task-1319166" [ 1078.385866] env[62521]: _type = "Task" [ 1078.385866] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.393753] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319166, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.473139] env[62521]: DEBUG nova.network.neutron [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Successfully created port: ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1078.623021] env[62521]: DEBUG nova.compute.manager [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1078.640483] env[62521]: DEBUG nova.network.neutron [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Updating instance_info_cache with network_info: [{"id": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "address": "fa:16:3e:44:95:f8", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee2404e8-7a", "ovs_interfaceid": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.808514] env[62521]: DEBUG nova.scheduler.client.report [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.823634] env[62521]: DEBUG nova.network.neutron [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1078.898993] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319166, 'name': RemoveSnapshot_Task, 'duration_secs': 0.370295} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.899309] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1078.899621] env[62521]: DEBUG nova.compute.manager [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.900439] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936d8002-f905-4d8e-bae2-b9abe11fc7cf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.977874] env[62521]: DEBUG nova.network.neutron [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Updating instance_info_cache with network_info: [{"id": "51638ad7-153a-49c6-8d83-8efc154d8502", "address": "fa:16:3e:b2:be:43", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51638ad7-15", "ovs_interfaceid": "51638ad7-153a-49c6-8d83-8efc154d8502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.142900] env[62521]: DEBUG oslo_concurrency.lockutils [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.143690] env[62521]: DEBUG nova.objects.instance [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lazy-loading 'flavor' on Instance uuid 703ff423-98b1-4942-b6fd-8b95fe57bd0a {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.313296] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.703s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.355749] env[62521]: INFO nova.scheduler.client.report [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Deleted allocations for instance d574f077-b221-4f91-8b54-0915421cb36f [ 1079.412791] env[62521]: INFO nova.compute.manager [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Shelve offloading [ 1079.414402] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1079.414770] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe730295-dc52-4aeb-b417-e4e03359334c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.422615] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1079.422615] env[62521]: value = "task-1319167" [ 1079.422615] env[62521]: _type = "Task" [ 1079.422615] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.431118] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.480518] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "refresh_cache-fbb2c755-3289-453d-ad19-40cc53b56fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.480870] env[62521]: DEBUG nova.compute.manager [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Instance network_info: |[{"id": "51638ad7-153a-49c6-8d83-8efc154d8502", "address": "fa:16:3e:b2:be:43", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51638ad7-15", "ovs_interfaceid": "51638ad7-153a-49c6-8d83-8efc154d8502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1079.481319] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:be:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '304be4f7-4e36-4468-9ef4-e457341cef18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51638ad7-153a-49c6-8d83-8efc154d8502', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1079.488862] env[62521]: DEBUG oslo.service.loopingcall [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.489086] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1079.489318] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-85b92f12-1099-44de-9c20-9c13f88bc782 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.509441] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1079.509441] env[62521]: value = "task-1319168" [ 1079.509441] env[62521]: _type = "Task" [ 1079.509441] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.517177] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319168, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.631583] env[62521]: DEBUG nova.compute.manager [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1079.653317] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590ecbef-7f58-4ad8-9f7a-db5d3aac2d5d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.657745] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1079.657979] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1079.658158] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1079.658347] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1079.658496] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1079.658674] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1079.658888] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1079.659061] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1079.659235] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1079.659400] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1079.659603] env[62521]: DEBUG nova.virt.hardware [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1079.660389] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360c85f9-0ef0-40d9-92e5-d8bdea1d909d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.682906] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1079.685060] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21bcb92c-e1d3-409e-85c1-074f30c6c26a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.687563] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3339b03-8bdc-44ad-afd8-37b36b49c7fb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.702389] env[62521]: DEBUG oslo_vmware.api [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1079.702389] env[62521]: value = "task-1319169" [ 1079.702389] env[62521]: _type = "Task" [ 1079.702389] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.709997] env[62521]: DEBUG oslo_vmware.api [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319169, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.712464] env[62521]: DEBUG nova.compute.manager [req-f8948e96-9387-4c9c-8180-b13f0cc3a40e req-c6920293-d38d-45d6-ae96-86a994b13012 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Received event network-changed-51638ad7-153a-49c6-8d83-8efc154d8502 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1079.712651] env[62521]: DEBUG nova.compute.manager [req-f8948e96-9387-4c9c-8180-b13f0cc3a40e req-c6920293-d38d-45d6-ae96-86a994b13012 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Refreshing instance network info cache due to event network-changed-51638ad7-153a-49c6-8d83-8efc154d8502. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1079.712861] env[62521]: DEBUG oslo_concurrency.lockutils [req-f8948e96-9387-4c9c-8180-b13f0cc3a40e req-c6920293-d38d-45d6-ae96-86a994b13012 service nova] Acquiring lock "refresh_cache-fbb2c755-3289-453d-ad19-40cc53b56fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.712998] env[62521]: DEBUG oslo_concurrency.lockutils [req-f8948e96-9387-4c9c-8180-b13f0cc3a40e req-c6920293-d38d-45d6-ae96-86a994b13012 service nova] Acquired lock "refresh_cache-fbb2c755-3289-453d-ad19-40cc53b56fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.713184] env[62521]: DEBUG nova.network.neutron [req-f8948e96-9387-4c9c-8180-b13f0cc3a40e req-c6920293-d38d-45d6-ae96-86a994b13012 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Refreshing network info cache for port 51638ad7-153a-49c6-8d83-8efc154d8502 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1079.863651] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5e377047-776b-4af9-bba2-d0cc3692bcdc tempest-ServersTestJSON-2132166461 tempest-ServersTestJSON-2132166461-project-member] Lock "d574f077-b221-4f91-8b54-0915421cb36f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.904s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.934940] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1079.935573] env[62521]: DEBUG nova.compute.manager [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1079.936644] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28d1e88-e30e-4df5-ad94-ad9a7a36abb7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.942747] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.942929] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.943122] env[62521]: DEBUG nova.network.neutron [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1079.968774] env[62521]: DEBUG nova.network.neutron [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Successfully updated port: ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1080.019058] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319168, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.212737] env[62521]: DEBUG oslo_vmware.api [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319169, 'name': PowerOffVM_Task, 'duration_secs': 0.23984} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.216012] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.218711] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1080.220993] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-908957cd-0a9d-4d4e-8469-7bd3ea7df07a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.242017] env[62521]: DEBUG oslo_vmware.api [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1080.242017] env[62521]: value = "task-1319170" [ 1080.242017] env[62521]: _type = "Task" [ 1080.242017] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.248139] env[62521]: DEBUG oslo_vmware.api [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319170, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.439605] env[62521]: DEBUG nova.network.neutron [req-f8948e96-9387-4c9c-8180-b13f0cc3a40e req-c6920293-d38d-45d6-ae96-86a994b13012 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Updated VIF entry in instance network info cache for port 51638ad7-153a-49c6-8d83-8efc154d8502. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1080.440180] env[62521]: DEBUG nova.network.neutron [req-f8948e96-9387-4c9c-8180-b13f0cc3a40e req-c6920293-d38d-45d6-ae96-86a994b13012 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Updating instance_info_cache with network_info: [{"id": "51638ad7-153a-49c6-8d83-8efc154d8502", "address": "fa:16:3e:b2:be:43", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51638ad7-15", "ovs_interfaceid": "51638ad7-153a-49c6-8d83-8efc154d8502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.471586] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.475020] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.475020] env[62521]: DEBUG nova.network.neutron [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1080.525528] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319168, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.667724] env[62521]: DEBUG nova.network.neutron [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating instance_info_cache with network_info: [{"id": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "address": "fa:16:3e:dd:37:93", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2f01aec-5d", "ovs_interfaceid": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.751092] env[62521]: DEBUG oslo_vmware.api [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319170, 'name': ReconfigVM_Task, 'duration_secs': 0.294161} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.751092] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1080.751092] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.751092] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b79ed05-2a35-4da7-9f62-1bd1088ca2f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.758377] env[62521]: DEBUG oslo_vmware.api [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1080.758377] env[62521]: value = "task-1319171" [ 1080.758377] env[62521]: _type = "Task" [ 1080.758377] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.766492] env[62521]: DEBUG oslo_vmware.api [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319171, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.942841] env[62521]: DEBUG oslo_concurrency.lockutils [req-f8948e96-9387-4c9c-8180-b13f0cc3a40e req-c6920293-d38d-45d6-ae96-86a994b13012 service nova] Releasing lock "refresh_cache-fbb2c755-3289-453d-ad19-40cc53b56fa2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.019147] env[62521]: DEBUG nova.network.neutron [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1081.024590] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319168, 'name': CreateVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.171288] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.195908] env[62521]: DEBUG nova.network.neutron [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance_info_cache with network_info: [{"id": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "address": "fa:16:3e:20:00:55", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff3e4ad4-29", "ovs_interfaceid": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.270393] env[62521]: DEBUG oslo_vmware.api [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319171, 'name': PowerOnVM_Task, 'duration_secs': 0.378296} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.270884] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.271149] env[62521]: DEBUG nova.compute.manager [None req-40b36ee8-beaf-4812-be2d-040e78adbec8 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1081.271918] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7140e338-2944-4a9e-b52b-29d5ef982278 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.526883] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319168, 'name': CreateVM_Task, 'duration_secs': 1.741025} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.527141] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1081.530810] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.530810] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.530810] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1081.530810] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5567c7e-933e-41c4-8a9c-01cbd7942838 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.537015] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1081.537015] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cefc14-3619-f62e-fcfa-6a5d3914af34" [ 1081.537015] env[62521]: _type = "Task" [ 1081.537015] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.545136] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cefc14-3619-f62e-fcfa-6a5d3914af34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.561856] env[62521]: DEBUG nova.compute.manager [req-70a9af6d-8ecb-48d9-a0a8-33e8291e71be req-3b84d7bf-a979-45ad-b1fd-667f9f4d34fb service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received event network-vif-unplugged-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.562158] env[62521]: DEBUG oslo_concurrency.lockutils [req-70a9af6d-8ecb-48d9-a0a8-33e8291e71be req-3b84d7bf-a979-45ad-b1fd-667f9f4d34fb service nova] Acquiring lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.562309] env[62521]: DEBUG oslo_concurrency.lockutils [req-70a9af6d-8ecb-48d9-a0a8-33e8291e71be req-3b84d7bf-a979-45ad-b1fd-667f9f4d34fb service nova] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.562528] env[62521]: DEBUG oslo_concurrency.lockutils [req-70a9af6d-8ecb-48d9-a0a8-33e8291e71be req-3b84d7bf-a979-45ad-b1fd-667f9f4d34fb service nova] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.562707] env[62521]: DEBUG nova.compute.manager [req-70a9af6d-8ecb-48d9-a0a8-33e8291e71be req-3b84d7bf-a979-45ad-b1fd-667f9f4d34fb service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] No waiting events found dispatching network-vif-unplugged-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1081.562902] env[62521]: WARNING nova.compute.manager [req-70a9af6d-8ecb-48d9-a0a8-33e8291e71be req-3b84d7bf-a979-45ad-b1fd-667f9f4d34fb service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received unexpected event network-vif-unplugged-a2f01aec-5d56-42ef-ac6c-6364f1571174 for instance with vm_state shelved and task_state shelving_offloading. [ 1081.640359] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1081.641318] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04234bcb-bc70-4d67-90e8-8fc17e5392a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.649180] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1081.649434] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-44fccc03-2166-49e8-855f-57d3aa7bc179 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.698198] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.698546] env[62521]: DEBUG nova.compute.manager [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Instance network_info: |[{"id": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "address": "fa:16:3e:20:00:55", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff3e4ad4-29", "ovs_interfaceid": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1081.699136] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:00:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff3e4ad4-29de-4ab1-bf26-bd8554ee6260', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1081.707986] env[62521]: DEBUG oslo.service.loopingcall [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.709362] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1081.709672] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1081.709879] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1081.710071] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleting the datastore file [datastore1] 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1081.710303] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed3d7b78-804a-4760-a515-aa31d4044640 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.724538] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f6e16b2-6a53-49da-97bd-7ad8828bb765 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.730549] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1081.730549] env[62521]: value = "task-1319173" [ 1081.730549] env[62521]: _type = "Task" [ 1081.730549] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.731921] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1081.731921] env[62521]: value = "task-1319174" [ 1081.731921] env[62521]: _type = "Task" [ 1081.731921] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.739909] env[62521]: DEBUG nova.compute.manager [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Received event network-vif-plugged-ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.740157] env[62521]: DEBUG oslo_concurrency.lockutils [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] Acquiring lock "e159807b-c7b0-4d2e-a7be-426ecaf01785-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.740363] env[62521]: DEBUG oslo_concurrency.lockutils [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.740533] env[62521]: DEBUG oslo_concurrency.lockutils [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.740702] env[62521]: DEBUG nova.compute.manager [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] No waiting events found dispatching network-vif-plugged-ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1081.740904] env[62521]: WARNING nova.compute.manager [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Received unexpected event network-vif-plugged-ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 for instance with vm_state building and task_state spawning. [ 1081.741091] env[62521]: DEBUG nova.compute.manager [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Received event network-changed-ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.741262] env[62521]: DEBUG nova.compute.manager [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Refreshing instance network info cache due to event network-changed-ff3e4ad4-29de-4ab1-bf26-bd8554ee6260. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1081.741451] env[62521]: DEBUG oslo_concurrency.lockutils [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] Acquiring lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.741593] env[62521]: DEBUG oslo_concurrency.lockutils [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] Acquired lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.741749] env[62521]: DEBUG nova.network.neutron [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Refreshing network info cache for port ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1081.746943] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319173, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.751388] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.046796] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cefc14-3619-f62e-fcfa-6a5d3914af34, 'name': SearchDatastore_Task, 'duration_secs': 0.010182} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.047163] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.047406] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1082.047645] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.047797] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.047983] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1082.048272] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31a7b8c8-0670-4cfd-9e65-733450dfa3a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.056737] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1082.056918] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1082.057663] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee5796b6-b724-4c4e-9b62-78e80d0d20f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.062951] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1082.062951] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52580aef-570c-080f-cf53-961fc7d7b578" [ 1082.062951] env[62521]: _type = "Task" [ 1082.062951] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.070349] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52580aef-570c-080f-cf53-961fc7d7b578, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.244892] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319173, 'name': CreateVM_Task, 'duration_secs': 0.319981} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.247652] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1082.249750] env[62521]: DEBUG oslo_vmware.api [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319174, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139461} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.250340] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.250500] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.250813] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1082.251089] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1082.251268] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1082.251438] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1082.253356] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa62d2a1-b2db-4603-ba72-436cc115d751 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.258249] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1082.258249] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5273037a-8d57-44d4-3ffd-5ac75f0648c9" [ 1082.258249] env[62521]: _type = "Task" [ 1082.258249] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.267745] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5273037a-8d57-44d4-3ffd-5ac75f0648c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.299432] env[62521]: INFO nova.scheduler.client.report [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleted allocations for instance 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 [ 1082.488120] env[62521]: DEBUG nova.network.neutron [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updated VIF entry in instance network info cache for port ff3e4ad4-29de-4ab1-bf26-bd8554ee6260. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1082.488484] env[62521]: DEBUG nova.network.neutron [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance_info_cache with network_info: [{"id": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "address": "fa:16:3e:20:00:55", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff3e4ad4-29", "ovs_interfaceid": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.573834] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52580aef-570c-080f-cf53-961fc7d7b578, 'name': SearchDatastore_Task, 'duration_secs': 0.011203} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.574704] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d93764b-a754-4b56-9892-f423899a4f04 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.580085] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1082.580085] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524d0656-34af-fd9a-c09c-26686b74ea8a" [ 1082.580085] env[62521]: _type = "Task" [ 1082.580085] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.589389] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524d0656-34af-fd9a-c09c-26686b74ea8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.768089] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5273037a-8d57-44d4-3ffd-5ac75f0648c9, 'name': SearchDatastore_Task, 'duration_secs': 0.009233} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.768389] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.769049] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1082.769049] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.807264] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.807503] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.807729] env[62521]: DEBUG nova.objects.instance [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'resources' on Instance uuid 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.991698] env[62521]: DEBUG oslo_concurrency.lockutils [req-89066514-e559-4b53-b1f4-5ef0fa68da21 req-aa27eed2-7b59-4b92-bb52-fcf4240eee39 service nova] Releasing lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.090732] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524d0656-34af-fd9a-c09c-26686b74ea8a, 'name': SearchDatastore_Task, 'duration_secs': 0.010582} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.091000] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.091289] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] fbb2c755-3289-453d-ad19-40cc53b56fa2/fbb2c755-3289-453d-ad19-40cc53b56fa2.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1083.091574] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.091768] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.091988] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-44822cb8-06ee-4bc8-abb8-abd0d95020e5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.094063] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75015db5-b00c-4524-aef6-d0c315594ef7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.101488] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1083.101488] env[62521]: value = "task-1319175" [ 1083.101488] env[62521]: _type = "Task" [ 1083.101488] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.102758] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.102758] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1083.106148] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca025abe-34f0-4722-93ff-81b692ea7268 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.115412] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319175, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.116455] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1083.116455] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e8fe9a-d388-e2c6-65b9-371834841f16" [ 1083.116455] env[62521]: _type = "Task" [ 1083.116455] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.125852] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e8fe9a-d388-e2c6-65b9-371834841f16, 'name': SearchDatastore_Task, 'duration_secs': 0.009243} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.126684] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87c13798-f44c-438c-a885-ca5bf28eac36 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.131995] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1083.131995] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5210e42b-7a19-aab4-dc39-f6374960e245" [ 1083.131995] env[62521]: _type = "Task" [ 1083.131995] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.140587] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5210e42b-7a19-aab4-dc39-f6374960e245, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.310805] env[62521]: DEBUG nova.objects.instance [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'numa_topology' on Instance uuid 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.593447] env[62521]: DEBUG nova.compute.manager [req-d9844692-469d-4074-a28e-c8e94c688352 req-bc3e84aa-d54a-4727-8d97-2ef98389e588 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received event network-changed-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.593651] env[62521]: DEBUG nova.compute.manager [req-d9844692-469d-4074-a28e-c8e94c688352 req-bc3e84aa-d54a-4727-8d97-2ef98389e588 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Refreshing instance network info cache due to event network-changed-a2f01aec-5d56-42ef-ac6c-6364f1571174. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1083.593846] env[62521]: DEBUG oslo_concurrency.lockutils [req-d9844692-469d-4074-a28e-c8e94c688352 req-bc3e84aa-d54a-4727-8d97-2ef98389e588 service nova] Acquiring lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.593993] env[62521]: DEBUG oslo_concurrency.lockutils [req-d9844692-469d-4074-a28e-c8e94c688352 req-bc3e84aa-d54a-4727-8d97-2ef98389e588 service nova] Acquired lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.594202] env[62521]: DEBUG nova.network.neutron [req-d9844692-469d-4074-a28e-c8e94c688352 req-bc3e84aa-d54a-4727-8d97-2ef98389e588 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Refreshing network info cache for port a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1083.611975] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319175, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497371} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.612272] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] fbb2c755-3289-453d-ad19-40cc53b56fa2/fbb2c755-3289-453d-ad19-40cc53b56fa2.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1083.612492] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1083.612785] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fdb7bfcb-893f-49eb-a70b-50814ff2f4cf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.618753] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1083.618753] env[62521]: value = "task-1319176" [ 1083.618753] env[62521]: _type = "Task" [ 1083.618753] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.626056] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319176, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.641046] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5210e42b-7a19-aab4-dc39-f6374960e245, 'name': SearchDatastore_Task, 'duration_secs': 0.008569} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.641323] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.641599] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] e159807b-c7b0-4d2e-a7be-426ecaf01785/e159807b-c7b0-4d2e-a7be-426ecaf01785.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1083.641838] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8019d3a3-4227-41e4-9b7d-9447a90deec3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.647650] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1083.647650] env[62521]: value = "task-1319177" [ 1083.647650] env[62521]: _type = "Task" [ 1083.647650] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.655697] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319177, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.814912] env[62521]: DEBUG nova.objects.base [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Object Instance<6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7> lazy-loaded attributes: resources,numa_topology {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1083.949147] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9d5e7c-5a78-49e6-81dd-57f4e461dd05 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.959388] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7273acb6-e699-4dee-8fc5-b9e12f52a139 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.998257] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2f24c7-89e8-4464-9ed6-f14816db5ba5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.006932] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83582c1-7a1f-40b0-997b-b5c6ceb5c82c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.021922] env[62521]: DEBUG nova.compute.provider_tree [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.129906] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319176, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065089} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.129906] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1084.130554] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7034081e-f05b-4119-97d0-4759db3ed242 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.157266] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] fbb2c755-3289-453d-ad19-40cc53b56fa2/fbb2c755-3289-453d-ad19-40cc53b56fa2.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.160036] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e73718ca-62a5-4ab4-9826-079cda21d569 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.184448] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319177, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472226} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.185657] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] e159807b-c7b0-4d2e-a7be-426ecaf01785/e159807b-c7b0-4d2e-a7be-426ecaf01785.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1084.185934] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1084.186293] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1084.186293] env[62521]: value = "task-1319178" [ 1084.186293] env[62521]: _type = "Task" [ 1084.186293] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.186487] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7d4506c3-2538-4278-a386-65df9aa93907 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.196249] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319178, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.197497] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1084.197497] env[62521]: value = "task-1319179" [ 1084.197497] env[62521]: _type = "Task" [ 1084.197497] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.207872] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319179, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.450117] env[62521]: DEBUG nova.network.neutron [req-d9844692-469d-4074-a28e-c8e94c688352 req-bc3e84aa-d54a-4727-8d97-2ef98389e588 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updated VIF entry in instance network info cache for port a2f01aec-5d56-42ef-ac6c-6364f1571174. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1084.450532] env[62521]: DEBUG nova.network.neutron [req-d9844692-469d-4074-a28e-c8e94c688352 req-bc3e84aa-d54a-4727-8d97-2ef98389e588 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating instance_info_cache with network_info: [{"id": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "address": "fa:16:3e:dd:37:93", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapa2f01aec-5d", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.524967] env[62521]: DEBUG nova.scheduler.client.report [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.697622] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319178, 'name': ReconfigVM_Task, 'duration_secs': 0.336962} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.697991] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Reconfigured VM instance instance-00000068 to attach disk [datastore1] fbb2c755-3289-453d-ad19-40cc53b56fa2/fbb2c755-3289-453d-ad19-40cc53b56fa2.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.701180] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-54506e82-27ab-452e-9cd9-525e4685f404 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.709306] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319179, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073598} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.710446] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1084.710837] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1084.710837] env[62521]: value = "task-1319180" [ 1084.710837] env[62521]: _type = "Task" [ 1084.710837] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.711404] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbdb471-43ac-41d4-8106-883112d13032 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.720787] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319180, 'name': Rename_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.737379] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] e159807b-c7b0-4d2e-a7be-426ecaf01785/e159807b-c7b0-4d2e-a7be-426ecaf01785.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.737623] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57da7101-482b-4d06-90de-c24e30136826 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.755930] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1084.755930] env[62521]: value = "task-1319181" [ 1084.755930] env[62521]: _type = "Task" [ 1084.755930] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.763260] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319181, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.953961] env[62521]: DEBUG oslo_concurrency.lockutils [req-d9844692-469d-4074-a28e-c8e94c688352 req-bc3e84aa-d54a-4727-8d97-2ef98389e588 service nova] Releasing lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.030285] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.223s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.073983] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.074266] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.133630] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.222678] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319180, 'name': Rename_Task, 'duration_secs': 0.134577} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.222924] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1085.223180] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7d497d4-3f51-4305-ba4c-e77762b257c1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.230774] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1085.230774] env[62521]: value = "task-1319182" [ 1085.230774] env[62521]: _type = "Task" [ 1085.230774] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.238351] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319182, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.265995] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319181, 'name': ReconfigVM_Task, 'duration_secs': 0.268081} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.266333] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Reconfigured VM instance instance-00000069 to attach disk [datastore1] e159807b-c7b0-4d2e-a7be-426ecaf01785/e159807b-c7b0-4d2e-a7be-426ecaf01785.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1085.267335] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d5859e9f-d750-4629-8bfa-af7e82236ec8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.272982] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1085.272982] env[62521]: value = "task-1319183" [ 1085.272982] env[62521]: _type = "Task" [ 1085.272982] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.282570] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319183, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.538397] env[62521]: DEBUG oslo_concurrency.lockutils [None req-3a853136-abea-4ed0-8cf4-d7e66c6ceb4e tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.176s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.538922] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.405s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.539268] env[62521]: INFO nova.compute.manager [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Unshelving [ 1085.576676] env[62521]: DEBUG nova.compute.manager [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1085.739991] env[62521]: DEBUG oslo_vmware.api [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319182, 'name': PowerOnVM_Task, 'duration_secs': 0.476626} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.740280] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1085.740488] env[62521]: INFO nova.compute.manager [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Took 8.40 seconds to spawn the instance on the hypervisor. [ 1085.740673] env[62521]: DEBUG nova.compute.manager [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1085.741485] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8ff833-5e90-4a30-b7a5-eb3c5bf76c0f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.782855] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319183, 'name': Rename_Task, 'duration_secs': 0.169355} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.783770] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1085.784033] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f8020d2-eafa-4e27-add7-d5e71d8aeaa4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.793118] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1085.793118] env[62521]: value = "task-1319184" [ 1085.793118] env[62521]: _type = "Task" [ 1085.793118] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.801457] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319184, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.099723] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.100241] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.101846] env[62521]: INFO nova.compute.claims [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1086.259233] env[62521]: INFO nova.compute.manager [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Took 13.18 seconds to build instance. [ 1086.302703] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319184, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.562369] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.761732] env[62521]: DEBUG oslo_concurrency.lockutils [None req-06b054c5-37d0-4c9e-a64f-5e2279e31d5b tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "fbb2c755-3289-453d-ad19-40cc53b56fa2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.693s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.803247] env[62521]: DEBUG oslo_vmware.api [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319184, 'name': PowerOnVM_Task, 'duration_secs': 0.767797} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.803452] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1086.803660] env[62521]: INFO nova.compute.manager [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Took 7.17 seconds to spawn the instance on the hypervisor. [ 1086.803842] env[62521]: DEBUG nova.compute.manager [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1086.804632] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704c65c2-b706-4bb0-a3e5-6c317e470590 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.970063] env[62521]: DEBUG oslo_concurrency.lockutils [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "fbb2c755-3289-453d-ad19-40cc53b56fa2" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.970063] env[62521]: DEBUG oslo_concurrency.lockutils [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "fbb2c755-3289-453d-ad19-40cc53b56fa2" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.970303] env[62521]: DEBUG nova.compute.manager [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1086.971138] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82af9269-0102-4d02-86d4-6f52aec9ecfc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.978538] env[62521]: DEBUG nova.compute.manager [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62521) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1086.979133] env[62521]: DEBUG nova.objects.instance [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lazy-loading 'flavor' on Instance uuid fbb2c755-3289-453d-ad19-40cc53b56fa2 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.226808] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5a3987-5fbb-4a68-b5be-1328804f265e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.234584] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d53d42-3919-4329-bec3-facd926778ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.265687] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c605059-5430-488c-92eb-097b3bb214af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.272902] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58175b06-0cb3-447d-90ee-8fa040cb1064 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.285582] env[62521]: DEBUG nova.compute.provider_tree [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.321071] env[62521]: INFO nova.compute.manager [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Took 14.18 seconds to build instance. [ 1087.484276] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1087.484276] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1c15016-a4fe-47b8-baa1-c568683c72a9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.491521] env[62521]: DEBUG oslo_vmware.api [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1087.491521] env[62521]: value = "task-1319185" [ 1087.491521] env[62521]: _type = "Task" [ 1087.491521] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.499970] env[62521]: DEBUG oslo_vmware.api [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319185, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.748025] env[62521]: DEBUG nova.compute.manager [req-3a047704-a87c-4d0e-acbf-8289a2028e2e req-a1a0e506-0229-48f8-9683-f100a09fa2c8 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Received event network-changed-ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.748246] env[62521]: DEBUG nova.compute.manager [req-3a047704-a87c-4d0e-acbf-8289a2028e2e req-a1a0e506-0229-48f8-9683-f100a09fa2c8 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Refreshing instance network info cache due to event network-changed-ff3e4ad4-29de-4ab1-bf26-bd8554ee6260. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1087.748463] env[62521]: DEBUG oslo_concurrency.lockutils [req-3a047704-a87c-4d0e-acbf-8289a2028e2e req-a1a0e506-0229-48f8-9683-f100a09fa2c8 service nova] Acquiring lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.748702] env[62521]: DEBUG oslo_concurrency.lockutils [req-3a047704-a87c-4d0e-acbf-8289a2028e2e req-a1a0e506-0229-48f8-9683-f100a09fa2c8 service nova] Acquired lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.748798] env[62521]: DEBUG nova.network.neutron [req-3a047704-a87c-4d0e-acbf-8289a2028e2e req-a1a0e506-0229-48f8-9683-f100a09fa2c8 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Refreshing network info cache for port ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1087.792021] env[62521]: DEBUG nova.scheduler.client.report [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.822959] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5a35d68b-d949-4563-bc5e-1fd6ad6ade60 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.690s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.001404] env[62521]: DEBUG oslo_vmware.api [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319185, 'name': PowerOffVM_Task, 'duration_secs': 0.21395} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.001690] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1088.001874] env[62521]: DEBUG nova.compute.manager [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1088.002631] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e09d73-448b-4b30-beb0-0bd38252cc12 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.294597] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.194s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.295290] env[62521]: DEBUG nova.compute.manager [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1088.298669] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.736s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.298812] env[62521]: DEBUG nova.objects.instance [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'pci_requests' on Instance uuid 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.486217] env[62521]: DEBUG nova.network.neutron [req-3a047704-a87c-4d0e-acbf-8289a2028e2e req-a1a0e506-0229-48f8-9683-f100a09fa2c8 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updated VIF entry in instance network info cache for port ff3e4ad4-29de-4ab1-bf26-bd8554ee6260. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1088.486662] env[62521]: DEBUG nova.network.neutron [req-3a047704-a87c-4d0e-acbf-8289a2028e2e req-a1a0e506-0229-48f8-9683-f100a09fa2c8 service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance_info_cache with network_info: [{"id": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "address": "fa:16:3e:20:00:55", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff3e4ad4-29", "ovs_interfaceid": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.513193] env[62521]: DEBUG oslo_concurrency.lockutils [None req-50e945ae-6949-4e5e-ba7f-aad8d00b3c1f tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "fbb2c755-3289-453d-ad19-40cc53b56fa2" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.543s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.802105] env[62521]: DEBUG nova.compute.utils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1088.802636] env[62521]: DEBUG nova.compute.manager [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1088.802794] env[62521]: DEBUG nova.network.neutron [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1088.805970] env[62521]: DEBUG nova.objects.instance [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'numa_topology' on Instance uuid 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.864154] env[62521]: DEBUG nova.policy [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5f378733a3043b3b66f005bc0dce756', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2d7ab38464a4c02a9dc94e04b4793ff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1088.989367] env[62521]: DEBUG oslo_concurrency.lockutils [req-3a047704-a87c-4d0e-acbf-8289a2028e2e req-a1a0e506-0229-48f8-9683-f100a09fa2c8 service nova] Releasing lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.186125] env[62521]: DEBUG nova.network.neutron [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Successfully created port: 02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1089.305726] env[62521]: DEBUG nova.compute.manager [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1089.309785] env[62521]: INFO nova.compute.claims [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1089.472764] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "fbb2c755-3289-453d-ad19-40cc53b56fa2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.472764] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "fbb2c755-3289-453d-ad19-40cc53b56fa2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.472764] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "fbb2c755-3289-453d-ad19-40cc53b56fa2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.473021] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "fbb2c755-3289-453d-ad19-40cc53b56fa2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.473288] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "fbb2c755-3289-453d-ad19-40cc53b56fa2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.475365] env[62521]: INFO nova.compute.manager [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Terminating instance [ 1089.477452] env[62521]: DEBUG nova.compute.manager [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1089.477562] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1089.478452] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b745473-0a68-44e4-951f-ed3c1fe3f078 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.487189] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1089.487438] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-595aff94-8d22-484e-9782-cad0ad7c8548 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.698096] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1089.698422] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1089.698563] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleting the datastore file [datastore1] fbb2c755-3289-453d-ad19-40cc53b56fa2 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.698961] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b4ad842-4359-488e-870d-8c3053cd28aa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.706233] env[62521]: DEBUG oslo_vmware.api [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1089.706233] env[62521]: value = "task-1319187" [ 1089.706233] env[62521]: _type = "Task" [ 1089.706233] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.714995] env[62521]: DEBUG oslo_vmware.api [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319187, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.216316] env[62521]: DEBUG oslo_vmware.api [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319187, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144551} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.217023] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1090.217023] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1090.217228] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1090.217410] env[62521]: INFO nova.compute.manager [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Took 0.74 seconds to destroy the instance on the hypervisor. [ 1090.217665] env[62521]: DEBUG oslo.service.loopingcall [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.217861] env[62521]: DEBUG nova.compute.manager [-] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1090.217964] env[62521]: DEBUG nova.network.neutron [-] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1090.334689] env[62521]: DEBUG nova.compute.manager [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1090.364683] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1090.364951] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1090.365138] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1090.365334] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1090.365486] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1090.365638] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1090.365849] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1090.366025] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1090.366208] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1090.366377] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1090.366553] env[62521]: DEBUG nova.virt.hardware [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1090.367432] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2beacda5-965b-4148-b374-af6829913a7a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.377509] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc64698-b726-4303-824d-e23c0b56dfa6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.476898] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f845f4-b22d-4709-b60d-1e9efacfeaa1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.485130] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54e7ae4-2c63-467c-83f6-7ae8b6ad0081 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.489641] env[62521]: DEBUG nova.compute.manager [req-782cb665-9319-436b-96da-654c6c6ea8da req-87be2678-8a1c-4f05-b558-2c6dff5d7b45 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Received event network-vif-deleted-51638ad7-153a-49c6-8d83-8efc154d8502 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.489843] env[62521]: INFO nova.compute.manager [req-782cb665-9319-436b-96da-654c6c6ea8da req-87be2678-8a1c-4f05-b558-2c6dff5d7b45 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Neutron deleted interface 51638ad7-153a-49c6-8d83-8efc154d8502; detaching it from the instance and deleting it from the info cache [ 1090.490030] env[62521]: DEBUG nova.network.neutron [req-782cb665-9319-436b-96da-654c6c6ea8da req-87be2678-8a1c-4f05-b558-2c6dff5d7b45 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.519232] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7d71fa-bef3-49c7-927a-5b18e811f146 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.527479] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd2a7df-97b1-4cc1-b16d-2fe816aed8ae {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.541563] env[62521]: DEBUG nova.compute.provider_tree [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1090.603494] env[62521]: DEBUG nova.compute.manager [req-618506d1-bfda-4c64-a099-21efcc3b61c4 req-403e3d34-c275-43eb-858b-81b1c2e00f83 service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Received event network-vif-plugged-02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.603717] env[62521]: DEBUG oslo_concurrency.lockutils [req-618506d1-bfda-4c64-a099-21efcc3b61c4 req-403e3d34-c275-43eb-858b-81b1c2e00f83 service nova] Acquiring lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.603943] env[62521]: DEBUG oslo_concurrency.lockutils [req-618506d1-bfda-4c64-a099-21efcc3b61c4 req-403e3d34-c275-43eb-858b-81b1c2e00f83 service nova] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.604122] env[62521]: DEBUG oslo_concurrency.lockutils [req-618506d1-bfda-4c64-a099-21efcc3b61c4 req-403e3d34-c275-43eb-858b-81b1c2e00f83 service nova] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.604294] env[62521]: DEBUG nova.compute.manager [req-618506d1-bfda-4c64-a099-21efcc3b61c4 req-403e3d34-c275-43eb-858b-81b1c2e00f83 service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] No waiting events found dispatching network-vif-plugged-02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1090.604461] env[62521]: WARNING nova.compute.manager [req-618506d1-bfda-4c64-a099-21efcc3b61c4 req-403e3d34-c275-43eb-858b-81b1c2e00f83 service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Received unexpected event network-vif-plugged-02aabc96-04f0-4b09-bcf6-44443d6542af for instance with vm_state building and task_state spawning. [ 1090.695308] env[62521]: DEBUG nova.network.neutron [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Successfully updated port: 02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1090.965094] env[62521]: DEBUG nova.network.neutron [-] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.992529] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf7f9e9f-ea90-41f3-8102-1e8f42f22235 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.004376] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e7a4c6-103e-495f-9810-022fe287a8d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.031490] env[62521]: DEBUG nova.compute.manager [req-782cb665-9319-436b-96da-654c6c6ea8da req-87be2678-8a1c-4f05-b558-2c6dff5d7b45 service nova] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Detach interface failed, port_id=51638ad7-153a-49c6-8d83-8efc154d8502, reason: Instance fbb2c755-3289-453d-ad19-40cc53b56fa2 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1091.044737] env[62521]: DEBUG nova.scheduler.client.report [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1091.197869] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.198050] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.198289] env[62521]: DEBUG nova.network.neutron [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1091.467868] env[62521]: INFO nova.compute.manager [-] [instance: fbb2c755-3289-453d-ad19-40cc53b56fa2] Took 1.25 seconds to deallocate network for instance. [ 1091.549763] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.251s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.580465] env[62521]: INFO nova.network.neutron [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating port a2f01aec-5d56-42ef-ac6c-6364f1571174 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1091.760773] env[62521]: DEBUG nova.network.neutron [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1091.973687] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.973954] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.974198] env[62521]: DEBUG nova.objects.instance [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lazy-loading 'resources' on Instance uuid fbb2c755-3289-453d-ad19-40cc53b56fa2 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.066400] env[62521]: DEBUG nova.network.neutron [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updating instance_info_cache with network_info: [{"id": "02aabc96-04f0-4b09-bcf6-44443d6542af", "address": "fa:16:3e:4a:64:63", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aabc96-04", "ovs_interfaceid": "02aabc96-04f0-4b09-bcf6-44443d6542af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.317561] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.317809] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.317983] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.318142] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1092.318307] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.318431] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Cleaning up deleted instances {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1092.569231] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.569542] env[62521]: DEBUG nova.compute.manager [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Instance network_info: |[{"id": "02aabc96-04f0-4b09-bcf6-44443d6542af", "address": "fa:16:3e:4a:64:63", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aabc96-04", "ovs_interfaceid": "02aabc96-04f0-4b09-bcf6-44443d6542af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1092.570150] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:64:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e4e52d8a-b086-4333-a5a1-938680a2d2bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02aabc96-04f0-4b09-bcf6-44443d6542af', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1092.578431] env[62521]: DEBUG oslo.service.loopingcall [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.580878] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1092.581310] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-259ce16c-e490-40d7-b2cb-e5b739725908 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.609231] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1092.609231] env[62521]: value = "task-1319188" [ 1092.609231] env[62521]: _type = "Task" [ 1092.609231] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.613874] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90bb1a4-bc92-4c79-bfba-8c1f88285a65 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.619066] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319188, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.623549] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c83eb88-7f4d-42ad-9f3d-d97a5b9e39ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.630779] env[62521]: DEBUG nova.compute.manager [req-4751ea64-59f3-4d4d-ac62-1a78ed15b083 req-72961ce8-25c7-4b73-84f1-4d9debe8b982 service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Received event network-changed-02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.630970] env[62521]: DEBUG nova.compute.manager [req-4751ea64-59f3-4d4d-ac62-1a78ed15b083 req-72961ce8-25c7-4b73-84f1-4d9debe8b982 service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Refreshing instance network info cache due to event network-changed-02aabc96-04f0-4b09-bcf6-44443d6542af. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1092.631296] env[62521]: DEBUG oslo_concurrency.lockutils [req-4751ea64-59f3-4d4d-ac62-1a78ed15b083 req-72961ce8-25c7-4b73-84f1-4d9debe8b982 service nova] Acquiring lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.631391] env[62521]: DEBUG oslo_concurrency.lockutils [req-4751ea64-59f3-4d4d-ac62-1a78ed15b083 req-72961ce8-25c7-4b73-84f1-4d9debe8b982 service nova] Acquired lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.631555] env[62521]: DEBUG nova.network.neutron [req-4751ea64-59f3-4d4d-ac62-1a78ed15b083 req-72961ce8-25c7-4b73-84f1-4d9debe8b982 service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Refreshing network info cache for port 02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1092.661179] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703d0779-8920-4a19-8420-4e596ee843d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.671800] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219c5e9c-ad2f-4abd-87d8-507ea24e558c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.687300] env[62521]: DEBUG nova.compute.provider_tree [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.831619] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] There are 45 instances to clean {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1092.831906] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: bc7d76b1-76b2-49a7-8188-0672312767a1] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.870558] env[62521]: DEBUG nova.network.neutron [req-4751ea64-59f3-4d4d-ac62-1a78ed15b083 req-72961ce8-25c7-4b73-84f1-4d9debe8b982 service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updated VIF entry in instance network info cache for port 02aabc96-04f0-4b09-bcf6-44443d6542af. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1092.870955] env[62521]: DEBUG nova.network.neutron [req-4751ea64-59f3-4d4d-ac62-1a78ed15b083 req-72961ce8-25c7-4b73-84f1-4d9debe8b982 service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updating instance_info_cache with network_info: [{"id": "02aabc96-04f0-4b09-bcf6-44443d6542af", "address": "fa:16:3e:4a:64:63", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aabc96-04", "ovs_interfaceid": "02aabc96-04f0-4b09-bcf6-44443d6542af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.077454] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.077662] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.077845] env[62521]: DEBUG nova.network.neutron [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1093.119228] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319188, 'name': CreateVM_Task, 'duration_secs': 0.302276} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.119392] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1093.120091] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.120265] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.120595] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1093.120848] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37093b7c-96df-4b4e-9b4b-f9d59eb11750 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.125521] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1093.125521] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526b2ed5-8ba3-e450-016c-4e6285667a84" [ 1093.125521] env[62521]: _type = "Task" [ 1093.125521] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.133205] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526b2ed5-8ba3-e450-016c-4e6285667a84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.190604] env[62521]: DEBUG nova.scheduler.client.report [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.336738] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 60c23400-7eb2-4049-8a3b-77599098e334] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.373942] env[62521]: DEBUG oslo_concurrency.lockutils [req-4751ea64-59f3-4d4d-ac62-1a78ed15b083 req-72961ce8-25c7-4b73-84f1-4d9debe8b982 service nova] Releasing lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.636522] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526b2ed5-8ba3-e450-016c-4e6285667a84, 'name': SearchDatastore_Task, 'duration_secs': 0.012706} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.636953] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.637217] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1093.637455] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.637604] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.638085] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1093.638085] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f0bc170-ddc7-4b36-98f2-38795171ced8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.645701] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1093.645887] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1093.646612] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-481c12f8-1173-4bc3-b81f-ebc23d9c417c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.651837] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1093.651837] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5239be92-e938-8ba8-66d5-322ca3dd3b00" [ 1093.651837] env[62521]: _type = "Task" [ 1093.651837] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.660861] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5239be92-e938-8ba8-66d5-322ca3dd3b00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.695321] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.721s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.715018] env[62521]: INFO nova.scheduler.client.report [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted allocations for instance fbb2c755-3289-453d-ad19-40cc53b56fa2 [ 1093.793704] env[62521]: DEBUG nova.network.neutron [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating instance_info_cache with network_info: [{"id": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "address": "fa:16:3e:dd:37:93", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2f01aec-5d", "ovs_interfaceid": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.840320] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 83fe58c9-920b-422d-be08-e5d53bf551d3] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.164168] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5239be92-e938-8ba8-66d5-322ca3dd3b00, 'name': SearchDatastore_Task, 'duration_secs': 0.009163} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.165160] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89766766-102c-4e53-a77f-e517c3fff5c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.170796] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1094.170796] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5203f9cb-1003-0cdd-67ad-bfc97bca8731" [ 1094.170796] env[62521]: _type = "Task" [ 1094.170796] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.178514] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5203f9cb-1003-0cdd-67ad-bfc97bca8731, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.223133] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b1db7b37-5fb8-4bca-b27d-2bdc70a9234e tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "fbb2c755-3289-453d-ad19-40cc53b56fa2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.750s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.296130] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.324670] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f30ab32855318b17d338951909adf9e5',container_format='bare',created_at=2024-10-20T00:39:05Z,direct_url=,disk_format='vmdk',id=2adddf14-344e-4adb-9e63-62c5a1ab8f27,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-926802237-shelved',owner='6ab001785c25414abbaa4bfe3a6a0bb4',properties=ImageMetaProps,protected=,size=31661568,status='active',tags=,updated_at=2024-10-20T00:39:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1094.324912] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1094.325094] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.325285] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1094.325438] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.325591] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1094.325819] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1094.326011] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1094.326201] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1094.326373] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1094.326560] env[62521]: DEBUG nova.virt.hardware [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.327451] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c7f0a4-986c-48c5-ba1a-09a683a60577 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.336472] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2145d37f-a848-45e9-ab05-e82a4045b6e5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.350417] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: aa266be6-2bf1-445f-a968-2c144d30ecfa] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.352557] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:37:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ec3f9e71-839a-429d-b211-d3dfc98ca4f6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a2f01aec-5d56-42ef-ac6c-6364f1571174', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1094.360358] env[62521]: DEBUG oslo.service.loopingcall [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.361073] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1094.361531] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a00b5893-4449-4647-8d6b-90cd5841db85 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.382289] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1094.382289] env[62521]: value = "task-1319189" [ 1094.382289] env[62521]: _type = "Task" [ 1094.382289] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.390784] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319189, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.663068] env[62521]: DEBUG nova.compute.manager [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received event network-vif-plugged-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1094.663359] env[62521]: DEBUG oslo_concurrency.lockutils [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] Acquiring lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.663594] env[62521]: DEBUG oslo_concurrency.lockutils [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.663783] env[62521]: DEBUG oslo_concurrency.lockutils [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.664083] env[62521]: DEBUG nova.compute.manager [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] No waiting events found dispatching network-vif-plugged-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1094.664264] env[62521]: WARNING nova.compute.manager [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received unexpected event network-vif-plugged-a2f01aec-5d56-42ef-ac6c-6364f1571174 for instance with vm_state shelved_offloaded and task_state spawning. [ 1094.664502] env[62521]: DEBUG nova.compute.manager [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received event network-changed-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1094.664695] env[62521]: DEBUG nova.compute.manager [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Refreshing instance network info cache due to event network-changed-a2f01aec-5d56-42ef-ac6c-6364f1571174. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1094.664883] env[62521]: DEBUG oslo_concurrency.lockutils [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] Acquiring lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.665035] env[62521]: DEBUG oslo_concurrency.lockutils [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] Acquired lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.665244] env[62521]: DEBUG nova.network.neutron [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Refreshing network info cache for port a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1094.682784] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5203f9cb-1003-0cdd-67ad-bfc97bca8731, 'name': SearchDatastore_Task, 'duration_secs': 0.019953} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.683182] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.683456] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 1bea570f-f2b2-4e9a-8dbb-418025402d3d/1bea570f-f2b2-4e9a-8dbb-418025402d3d.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1094.683733] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a3fbac2-0aff-448d-a3c9-eb7dba8d4d9e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.693173] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1094.693173] env[62521]: value = "task-1319190" [ 1094.693173] env[62521]: _type = "Task" [ 1094.693173] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.701676] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.862497] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 31f926a8-777d-4882-afd6-1e4dba3f4e11] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.893369] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319189, 'name': CreateVM_Task, 'duration_secs': 0.32566} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.893560] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1094.894387] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.894568] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.895028] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1094.895384] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45e89667-c72f-42bf-afdc-b684ff995e62 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.901171] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1094.901171] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a78f8d-3724-d32f-c374-4c32e8f4d6b1" [ 1094.901171] env[62521]: _type = "Task" [ 1094.901171] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.911379] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a78f8d-3724-d32f-c374-4c32e8f4d6b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.206416] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319190, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.367803] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 69c5ff30-259d-4067-b5e2-ab727009d8ad] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.415151] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.415436] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Processing image 2adddf14-344e-4adb-9e63-62c5a1ab8f27 {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1095.415679] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27/2adddf14-344e-4adb-9e63-62c5a1ab8f27.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.415832] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27/2adddf14-344e-4adb-9e63-62c5a1ab8f27.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.416030] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1095.417392] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ae30f16-e39a-4e2a-9494-bd83dba62678 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.428263] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1095.428413] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1095.429264] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c31df50-dbc1-4a89-9c14-6600b3ec2ea8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.436276] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1095.436276] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521315ba-b896-653e-e3a6-c96f31b24b97" [ 1095.436276] env[62521]: _type = "Task" [ 1095.436276] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.437243] env[62521]: DEBUG nova.network.neutron [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updated VIF entry in instance network info cache for port a2f01aec-5d56-42ef-ac6c-6364f1571174. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1095.437580] env[62521]: DEBUG nova.network.neutron [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating instance_info_cache with network_info: [{"id": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "address": "fa:16:3e:dd:37:93", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2f01aec-5d", "ovs_interfaceid": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.447185] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]521315ba-b896-653e-e3a6-c96f31b24b97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.458196] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "e78762d1-a443-443e-ae8c-ab7aac7874bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.458374] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "e78762d1-a443-443e-ae8c-ab7aac7874bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.704883] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319190, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.615221} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.705322] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 1bea570f-f2b2-4e9a-8dbb-418025402d3d/1bea570f-f2b2-4e9a-8dbb-418025402d3d.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1095.705411] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1095.705626] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b70c0882-b075-4113-9539-ce3d1f4abb96 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.713517] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1095.713517] env[62521]: value = "task-1319191" [ 1095.713517] env[62521]: _type = "Task" [ 1095.713517] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.722407] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319191, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.870877] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7cdeed76-e5b3-4959-b6a6-610d23233e1b] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.943729] env[62521]: DEBUG oslo_concurrency.lockutils [req-e66c2cf2-2e01-42e3-af2f-0b890fde2448 req-3f266cfb-5e9b-4d6f-bd5c-dcab27e11916 service nova] Releasing lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.949985] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Preparing fetch location {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1095.950261] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Fetch image to [datastore2] OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f/OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f.vmdk {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1095.950450] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Downloading stream optimized image 2adddf14-344e-4adb-9e63-62c5a1ab8f27 to [datastore2] OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f/OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f.vmdk on the data store datastore2 as vApp {{(pid=62521) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1095.950624] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Downloading image file data 2adddf14-344e-4adb-9e63-62c5a1ab8f27 to the ESX as VM named 'OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f' {{(pid=62521) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1095.960294] env[62521]: DEBUG nova.compute.manager [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1096.028216] env[62521]: DEBUG oslo_vmware.rw_handles [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1096.028216] env[62521]: value = "resgroup-9" [ 1096.028216] env[62521]: _type = "ResourcePool" [ 1096.028216] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1096.028802] env[62521]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-33b842b0-5057-4c68-9d0a-e8acafce2c51 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.051810] env[62521]: DEBUG oslo_vmware.rw_handles [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lease: (returnval){ [ 1096.051810] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cf4820-e056-6d82-bd48-d201f72ccf4b" [ 1096.051810] env[62521]: _type = "HttpNfcLease" [ 1096.051810] env[62521]: } obtained for vApp import into resource pool (val){ [ 1096.051810] env[62521]: value = "resgroup-9" [ 1096.051810] env[62521]: _type = "ResourcePool" [ 1096.051810] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1096.052291] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the lease: (returnval){ [ 1096.052291] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cf4820-e056-6d82-bd48-d201f72ccf4b" [ 1096.052291] env[62521]: _type = "HttpNfcLease" [ 1096.052291] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1096.059031] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1096.059031] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cf4820-e056-6d82-bd48-d201f72ccf4b" [ 1096.059031] env[62521]: _type = "HttpNfcLease" [ 1096.059031] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1096.224014] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319191, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.157223} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.224272] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1096.225094] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc0ecc54-6b1d-4dad-8620-ccbf1e62e1c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.248013] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 1bea570f-f2b2-4e9a-8dbb-418025402d3d/1bea570f-f2b2-4e9a-8dbb-418025402d3d.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1096.248337] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e10786b-e7be-41fa-9a80-9c120c311913 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.268073] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1096.268073] env[62521]: value = "task-1319193" [ 1096.268073] env[62521]: _type = "Task" [ 1096.268073] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.276278] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319193, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.374348] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: a0646a00-1f1c-4892-bb32-6212c90e9e95] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.485121] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.485413] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.487480] env[62521]: INFO nova.compute.claims [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1096.561832] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1096.561832] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cf4820-e056-6d82-bd48-d201f72ccf4b" [ 1096.561832] env[62521]: _type = "HttpNfcLease" [ 1096.561832] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1096.778893] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319193, 'name': ReconfigVM_Task, 'duration_secs': 0.275759} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.779244] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 1bea570f-f2b2-4e9a-8dbb-418025402d3d/1bea570f-f2b2-4e9a-8dbb-418025402d3d.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1096.779881] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b67719d7-a98f-45ad-a366-3a3d549188bc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.785854] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1096.785854] env[62521]: value = "task-1319194" [ 1096.785854] env[62521]: _type = "Task" [ 1096.785854] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.793811] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319194, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.878127] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: a68d4cb7-bfad-48f6-8aeb-bca04248a0d7] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.061647] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1097.061647] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cf4820-e056-6d82-bd48-d201f72ccf4b" [ 1097.061647] env[62521]: _type = "HttpNfcLease" [ 1097.061647] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1097.061983] env[62521]: DEBUG oslo_vmware.rw_handles [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1097.061983] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52cf4820-e056-6d82-bd48-d201f72ccf4b" [ 1097.061983] env[62521]: _type = "HttpNfcLease" [ 1097.061983] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1097.062735] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8818c93-5f0f-4deb-9131-09e7994606e5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.071489] env[62521]: DEBUG oslo_vmware.rw_handles [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b9e5d-208c-ba78-41bc-897cb32300f1/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1097.071676] env[62521]: DEBUG oslo_vmware.rw_handles [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating HTTP connection to write to file with size = 31661568 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b9e5d-208c-ba78-41bc-897cb32300f1/disk-0.vmdk. {{(pid=62521) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1097.136919] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7a243dc4-f74e-427f-8f3d-e8911eca7da1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.296801] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319194, 'name': Rename_Task, 'duration_secs': 0.139535} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.297078] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1097.297336] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4b4e59b-9dc4-41ce-b2d5-f96b698d7e70 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.304831] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1097.304831] env[62521]: value = "task-1319195" [ 1097.304831] env[62521]: _type = "Task" [ 1097.304831] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.312371] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319195, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.382124] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7fa72dbb-d9c6-4e83-a30b-a7c2ac92615c] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.759168] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6fedac-19f5-4c78-bd92-ac79e2a9e35d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.768210] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca6e262-a8c0-43bd-b20f-9d2b18033337 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.807658] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f515555-00c1-4481-8ac5-82de6f8e9f6a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.826840] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24cd475b-d547-44d5-91db-b92eff0b6f90 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.831103] env[62521]: DEBUG oslo_vmware.api [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319195, 'name': PowerOnVM_Task, 'duration_secs': 0.511519} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.833088] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1097.833280] env[62521]: INFO nova.compute.manager [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Took 7.50 seconds to spawn the instance on the hypervisor. [ 1097.833634] env[62521]: DEBUG nova.compute.manager [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1097.834744] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f72bec3-e6df-4c9b-a275-bedd3ff56c13 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.846422] env[62521]: DEBUG nova.compute.provider_tree [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1097.885643] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: f90f2af0-1a83-4ce4-8ebb-2cbc699cc39d] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.245040] env[62521]: DEBUG oslo_vmware.rw_handles [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Completed reading data from the image iterator. {{(pid=62521) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1098.245040] env[62521]: DEBUG oslo_vmware.rw_handles [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b9e5d-208c-ba78-41bc-897cb32300f1/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1098.245869] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6e39d4-9a73-4f72-86dc-bc1eedd34153 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.252808] env[62521]: DEBUG oslo_vmware.rw_handles [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b9e5d-208c-ba78-41bc-897cb32300f1/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1098.253056] env[62521]: DEBUG oslo_vmware.rw_handles [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b9e5d-208c-ba78-41bc-897cb32300f1/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1098.253234] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-aeed4335-d732-4e07-86ed-a9ab6dcc293c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.368549] env[62521]: INFO nova.compute.manager [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Took 12.29 seconds to build instance. [ 1098.371942] env[62521]: ERROR nova.scheduler.client.report [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [req-2976934a-2714-42fd-bd02-be220cd27da8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2976934a-2714-42fd-bd02-be220cd27da8"}]} [ 1098.390645] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: b8158051-98b6-47d8-84c5-d1d56515afe8] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.416116] env[62521]: DEBUG nova.scheduler.client.report [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1098.465879] env[62521]: DEBUG nova.scheduler.client.report [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1098.466168] env[62521]: DEBUG nova.compute.provider_tree [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1098.494187] env[62521]: DEBUG oslo_vmware.rw_handles [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b9e5d-208c-ba78-41bc-897cb32300f1/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1098.494435] env[62521]: INFO nova.virt.vmwareapi.images [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Downloaded image file data 2adddf14-344e-4adb-9e63-62c5a1ab8f27 [ 1098.499899] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea763c3-e658-446d-8e14-0e88c899b438 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.541325] env[62521]: DEBUG nova.scheduler.client.report [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1098.543289] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5108b2eb-9ec9-48cb-abea-8b2b278e9cc3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.562114] env[62521]: DEBUG nova.scheduler.client.report [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1098.565866] env[62521]: INFO nova.virt.vmwareapi.images [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] The imported VM was unregistered [ 1098.568344] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Caching image {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1098.568573] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Creating directory with path [datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27 {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1098.568829] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc519686-c494-4f6a-805c-b07eaa04a8fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.588094] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Created directory with path [datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27 {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1098.588308] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f/OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f.vmdk to [datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27/2adddf14-344e-4adb-9e63-62c5a1ab8f27.vmdk. {{(pid=62521) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1098.588560] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-cfc6c36b-ff2b-4ab6-aa11-a4909f3f6d66 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.596954] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1098.596954] env[62521]: value = "task-1319197" [ 1098.596954] env[62521]: _type = "Task" [ 1098.596954] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.605917] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319197, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.677741] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92eded96-19f1-4869-b47c-06aea2a94480 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.685455] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5c92dc-0152-4012-af0b-3d3bbe709750 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.714704] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70e9ebf-c903-489e-a3c2-91c811dd6a9c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.721639] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40074734-fe52-4569-8bb5-1e331c86aa9a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.735577] env[62521]: DEBUG nova.compute.provider_tree [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1098.870628] env[62521]: DEBUG oslo_concurrency.lockutils [None req-be4bb421-5ba6-4fad-85b4-7e435f40eba3 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.796s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.922257] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 0743a5b6-87fe-4612-a5a1-5d6527b69eb2] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.106584] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319197, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.266831] env[62521]: DEBUG nova.scheduler.client.report [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 146 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1099.267259] env[62521]: DEBUG nova.compute.provider_tree [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 146 to 147 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1099.267473] env[62521]: DEBUG nova.compute.provider_tree [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1099.426604] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: ac420a77-2160-45d6-b313-c5304536c39d] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.606959] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319197, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.625439] env[62521]: DEBUG nova.compute.manager [req-1e11353e-4fb7-4222-a116-57bb9a25c554 req-27015ec3-bbb6-4b4a-a422-b37471e1098c service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Received event network-changed-02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1099.626611] env[62521]: DEBUG nova.compute.manager [req-1e11353e-4fb7-4222-a116-57bb9a25c554 req-27015ec3-bbb6-4b4a-a422-b37471e1098c service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Refreshing instance network info cache due to event network-changed-02aabc96-04f0-4b09-bcf6-44443d6542af. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1099.626611] env[62521]: DEBUG oslo_concurrency.lockutils [req-1e11353e-4fb7-4222-a116-57bb9a25c554 req-27015ec3-bbb6-4b4a-a422-b37471e1098c service nova] Acquiring lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.626611] env[62521]: DEBUG oslo_concurrency.lockutils [req-1e11353e-4fb7-4222-a116-57bb9a25c554 req-27015ec3-bbb6-4b4a-a422-b37471e1098c service nova] Acquired lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.626611] env[62521]: DEBUG nova.network.neutron [req-1e11353e-4fb7-4222-a116-57bb9a25c554 req-27015ec3-bbb6-4b4a-a422-b37471e1098c service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Refreshing network info cache for port 02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1099.772910] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.287s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.773483] env[62521]: DEBUG nova.compute.manager [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1099.930640] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 02c0ba3d-377b-4f66-9b1e-8cedc8d3fb4c] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.108719] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319197, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.278995] env[62521]: DEBUG nova.compute.utils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1100.282881] env[62521]: DEBUG nova.compute.manager [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1100.283072] env[62521]: DEBUG nova.network.neutron [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1100.325181] env[62521]: DEBUG nova.policy [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3566601952144168b5d400c41026c81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0dfb5f5bd4646ec95b7c6a2d0434175', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1100.394919] env[62521]: DEBUG nova.network.neutron [req-1e11353e-4fb7-4222-a116-57bb9a25c554 req-27015ec3-bbb6-4b4a-a422-b37471e1098c service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updated VIF entry in instance network info cache for port 02aabc96-04f0-4b09-bcf6-44443d6542af. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1100.395370] env[62521]: DEBUG nova.network.neutron [req-1e11353e-4fb7-4222-a116-57bb9a25c554 req-27015ec3-bbb6-4b4a-a422-b37471e1098c service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updating instance_info_cache with network_info: [{"id": "02aabc96-04f0-4b09-bcf6-44443d6542af", "address": "fa:16:3e:4a:64:63", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aabc96-04", "ovs_interfaceid": "02aabc96-04f0-4b09-bcf6-44443d6542af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.434209] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: cc4b29ee-969d-49c6-9fae-71d5f850aae1] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.581704] env[62521]: DEBUG nova.network.neutron [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Successfully created port: 13d38661-4ce7-48de-a860-47d294c7548e {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1100.610690] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319197, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.785850] env[62521]: DEBUG nova.compute.manager [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1100.898708] env[62521]: DEBUG oslo_concurrency.lockutils [req-1e11353e-4fb7-4222-a116-57bb9a25c554 req-27015ec3-bbb6-4b4a-a422-b37471e1098c service nova] Releasing lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.937744] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 3afba9cd-f66e-44c5-b660-b4314f370901] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.109602] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319197, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.443628] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 17876bcc-3a7b-47b0-afd9-1215e24864ae] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.515145] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1dbc7046-53ea-4a38-b644-21524cd42a5e tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.515416] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1dbc7046-53ea-4a38-b644-21524cd42a5e tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.515753] env[62521]: DEBUG nova.objects.instance [None req-1dbc7046-53ea-4a38-b644-21524cd42a5e tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'flavor' on Instance uuid 7cb9fe57-6940-4f84-bdde-32f185e28451 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.610913] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319197, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.67625} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.611212] env[62521]: INFO nova.virt.vmwareapi.ds_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f/OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f.vmdk to [datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27/2adddf14-344e-4adb-9e63-62c5a1ab8f27.vmdk. [ 1101.611420] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Cleaning up location [datastore2] OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1101.611590] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_c1803ea4-acec-4370-9f0e-9a951b2d158f {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1101.611843] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9212d68a-89f7-4c91-ab7b-825c3926a9f0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.618247] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1101.618247] env[62521]: value = "task-1319198" [ 1101.618247] env[62521]: _type = "Task" [ 1101.618247] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.625953] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319198, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.795698] env[62521]: DEBUG nova.compute.manager [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1101.820617] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1101.820852] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1101.821026] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1101.821223] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1101.821409] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1101.821567] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1101.821778] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1101.821942] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1101.822144] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1101.822394] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1101.822589] env[62521]: DEBUG nova.virt.hardware [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1101.823448] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da50783-594f-4cd3-8c4c-d586918aa2ee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.831431] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77747389-d174-46dd-93d8-75962e1a1827 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.947437] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 611462d2-7f57-44b0-ac36-db32af3d0dd0] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.992572] env[62521]: DEBUG nova.compute.manager [req-a8ce244f-c802-4e5c-ad05-8740a649b83e req-744a4abc-4215-4686-b333-514d12e08c22 service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Received event network-vif-plugged-13d38661-4ce7-48de-a860-47d294c7548e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.992572] env[62521]: DEBUG oslo_concurrency.lockutils [req-a8ce244f-c802-4e5c-ad05-8740a649b83e req-744a4abc-4215-4686-b333-514d12e08c22 service nova] Acquiring lock "e78762d1-a443-443e-ae8c-ab7aac7874bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.992572] env[62521]: DEBUG oslo_concurrency.lockutils [req-a8ce244f-c802-4e5c-ad05-8740a649b83e req-744a4abc-4215-4686-b333-514d12e08c22 service nova] Lock "e78762d1-a443-443e-ae8c-ab7aac7874bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.992572] env[62521]: DEBUG oslo_concurrency.lockutils [req-a8ce244f-c802-4e5c-ad05-8740a649b83e req-744a4abc-4215-4686-b333-514d12e08c22 service nova] Lock "e78762d1-a443-443e-ae8c-ab7aac7874bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.992736] env[62521]: DEBUG nova.compute.manager [req-a8ce244f-c802-4e5c-ad05-8740a649b83e req-744a4abc-4215-4686-b333-514d12e08c22 service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] No waiting events found dispatching network-vif-plugged-13d38661-4ce7-48de-a860-47d294c7548e {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1101.993470] env[62521]: WARNING nova.compute.manager [req-a8ce244f-c802-4e5c-ad05-8740a649b83e req-744a4abc-4215-4686-b333-514d12e08c22 service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Received unexpected event network-vif-plugged-13d38661-4ce7-48de-a860-47d294c7548e for instance with vm_state building and task_state spawning. [ 1102.019806] env[62521]: DEBUG nova.objects.instance [None req-1dbc7046-53ea-4a38-b644-21524cd42a5e tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'pci_requests' on Instance uuid 7cb9fe57-6940-4f84-bdde-32f185e28451 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.076285] env[62521]: DEBUG nova.network.neutron [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Successfully updated port: 13d38661-4ce7-48de-a860-47d294c7548e {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1102.128825] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319198, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.077957} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.129096] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1102.129247] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27/2adddf14-344e-4adb-9e63-62c5a1ab8f27.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.129494] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27/2adddf14-344e-4adb-9e63-62c5a1ab8f27.vmdk to [datastore2] 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7/6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1102.129746] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ddf8f3a9-471d-4d2f-8428-11ef55c254fe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.136055] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1102.136055] env[62521]: value = "task-1319199" [ 1102.136055] env[62521]: _type = "Task" [ 1102.136055] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.143613] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319199, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.450847] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 563ad3b2-e53e-4682-9cde-bd6f709718de] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.522818] env[62521]: DEBUG nova.objects.base [None req-1dbc7046-53ea-4a38-b644-21524cd42a5e tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Object Instance<7cb9fe57-6940-4f84-bdde-32f185e28451> lazy-loaded attributes: flavor,pci_requests {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1102.523107] env[62521]: DEBUG nova.network.neutron [None req-1dbc7046-53ea-4a38-b644-21524cd42a5e tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1102.580074] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "refresh_cache-e78762d1-a443-443e-ae8c-ab7aac7874bb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.580384] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "refresh_cache-e78762d1-a443-443e-ae8c-ab7aac7874bb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.580521] env[62521]: DEBUG nova.network.neutron [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1102.625074] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1dbc7046-53ea-4a38-b644-21524cd42a5e tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.109s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.647484] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319199, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.954955] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 11e61e72-a311-4c43-bf53-df82ad9d70f7] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.979747] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.979999] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.119514] env[62521]: DEBUG nova.network.neutron [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1103.149224] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319199, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.267469] env[62521]: DEBUG nova.network.neutron [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Updating instance_info_cache with network_info: [{"id": "13d38661-4ce7-48de-a860-47d294c7548e", "address": "fa:16:3e:9b:77:01", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13d38661-4c", "ovs_interfaceid": "13d38661-4ce7-48de-a860-47d294c7548e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.458686] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 6c4c36b7-b242-4578-9c23-d2529f308cb1] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.483234] env[62521]: DEBUG nova.compute.utils [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1103.650147] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319199, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.770261] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "refresh_cache-e78762d1-a443-443e-ae8c-ab7aac7874bb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.770610] env[62521]: DEBUG nova.compute.manager [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Instance network_info: |[{"id": "13d38661-4ce7-48de-a860-47d294c7548e", "address": "fa:16:3e:9b:77:01", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13d38661-4c", "ovs_interfaceid": "13d38661-4ce7-48de-a860-47d294c7548e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1103.771073] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:77:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '304be4f7-4e36-4468-9ef4-e457341cef18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '13d38661-4ce7-48de-a860-47d294c7548e', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1103.779224] env[62521]: DEBUG oslo.service.loopingcall [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.779545] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1103.779790] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fe41e25-9ac9-498e-8e2d-34ec412bc89a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.803285] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1103.803285] env[62521]: value = "task-1319200" [ 1103.803285] env[62521]: _type = "Task" [ 1103.803285] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.811561] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319200, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.962088] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: e2104c11-7713-4542-83da-bc4d2534113a] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.986745] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.150350] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319199, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.235176] env[62521]: DEBUG nova.compute.manager [req-cf7ccfad-4a4e-4eec-9795-96a076bef74d req-282da598-6e29-416b-98dc-16eedf46f538 service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Received event network-changed-13d38661-4ce7-48de-a860-47d294c7548e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.235447] env[62521]: DEBUG nova.compute.manager [req-cf7ccfad-4a4e-4eec-9795-96a076bef74d req-282da598-6e29-416b-98dc-16eedf46f538 service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Refreshing instance network info cache due to event network-changed-13d38661-4ce7-48de-a860-47d294c7548e. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1104.235679] env[62521]: DEBUG oslo_concurrency.lockutils [req-cf7ccfad-4a4e-4eec-9795-96a076bef74d req-282da598-6e29-416b-98dc-16eedf46f538 service nova] Acquiring lock "refresh_cache-e78762d1-a443-443e-ae8c-ab7aac7874bb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.235906] env[62521]: DEBUG oslo_concurrency.lockutils [req-cf7ccfad-4a4e-4eec-9795-96a076bef74d req-282da598-6e29-416b-98dc-16eedf46f538 service nova] Acquired lock "refresh_cache-e78762d1-a443-443e-ae8c-ab7aac7874bb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.236164] env[62521]: DEBUG nova.network.neutron [req-cf7ccfad-4a4e-4eec-9795-96a076bef74d req-282da598-6e29-416b-98dc-16eedf46f538 service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Refreshing network info cache for port 13d38661-4ce7-48de-a860-47d294c7548e {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1104.314285] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319200, 'name': CreateVM_Task, 'duration_secs': 0.458276} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.314447] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1104.315186] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.315364] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.315763] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1104.316033] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb9521c5-671b-4b29-81be-17ff18f8513d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.322261] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1104.322261] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52881878-c60f-15a2-b45e-42fa92c6e710" [ 1104.322261] env[62521]: _type = "Task" [ 1104.322261] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.330603] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52881878-c60f-15a2-b45e-42fa92c6e710, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.465506] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 3447712d-daa6-4329-8015-6474cca1a5e4] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.649971] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319199, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.253131} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.650265] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2adddf14-344e-4adb-9e63-62c5a1ab8f27/2adddf14-344e-4adb-9e63-62c5a1ab8f27.vmdk to [datastore2] 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7/6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1104.651025] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e8d36c-6fd6-4caf-959b-0f38821e1e0e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.671934] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7/6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7.vmdk or device None with type streamOptimized {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.672168] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8923c262-a371-4b0c-b895-ae23d2f856c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.690188] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1104.690188] env[62521]: value = "task-1319201" [ 1104.690188] env[62521]: _type = "Task" [ 1104.690188] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.697351] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319201, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.769526] env[62521]: DEBUG oslo_concurrency.lockutils [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.769848] env[62521]: DEBUG oslo_concurrency.lockutils [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.770169] env[62521]: DEBUG nova.objects.instance [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'flavor' on Instance uuid 7cb9fe57-6940-4f84-bdde-32f185e28451 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.831962] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52881878-c60f-15a2-b45e-42fa92c6e710, 'name': SearchDatastore_Task, 'duration_secs': 0.014783} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.834090] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.834331] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1104.834564] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.834711] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.834894] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1104.835179] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3fb2f2e4-6429-4b8d-b417-14b66a475d55 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.843390] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1104.843677] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1104.844597] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02dbd58b-ce48-44f9-a881-a80957daf994 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.851426] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1104.851426] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52074742-47dc-0378-b1e0-7677e2d52533" [ 1104.851426] env[62521]: _type = "Task" [ 1104.851426] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.863165] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52074742-47dc-0378-b1e0-7677e2d52533, 'name': SearchDatastore_Task, 'duration_secs': 0.009422} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.863934] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d1f9ca2-fb0b-442e-8921-fea3ee05586f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.869148] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1104.869148] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529ceacd-9124-2de2-6bbb-fd0a4dc54efc" [ 1104.869148] env[62521]: _type = "Task" [ 1104.869148] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.876401] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529ceacd-9124-2de2-6bbb-fd0a4dc54efc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.963221] env[62521]: DEBUG nova.network.neutron [req-cf7ccfad-4a4e-4eec-9795-96a076bef74d req-282da598-6e29-416b-98dc-16eedf46f538 service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Updated VIF entry in instance network info cache for port 13d38661-4ce7-48de-a860-47d294c7548e. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1104.963624] env[62521]: DEBUG nova.network.neutron [req-cf7ccfad-4a4e-4eec-9795-96a076bef74d req-282da598-6e29-416b-98dc-16eedf46f538 service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Updating instance_info_cache with network_info: [{"id": "13d38661-4ce7-48de-a860-47d294c7548e", "address": "fa:16:3e:9b:77:01", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13d38661-4c", "ovs_interfaceid": "13d38661-4ce7-48de-a860-47d294c7548e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.969698] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 1d1d2235-748a-4f27-a915-b69f9110257b] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.047408] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.047670] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.047909] env[62521]: INFO nova.compute.manager [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Attaching volume 0796a138-bf1a-4af3-82d5-800c1f80cba7 to /dev/sdb [ 1105.078483] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612f1d62-76bc-49c3-ba3b-3a7ed0bf91cd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.085548] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60109ddc-7bb0-45a4-8ffa-bb873bd11541 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.098777] env[62521]: DEBUG nova.virt.block_device [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating existing volume attachment record: 136db45d-cd65-4521-9051-fa8078b7edb4 {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1105.200019] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319201, 'name': ReconfigVM_Task, 'duration_secs': 0.294118} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.200199] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7/6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7.vmdk or device None with type streamOptimized {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.200830] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94e3d1ba-88a8-4beb-b63a-02ccded67ce9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.206376] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1105.206376] env[62521]: value = "task-1319202" [ 1105.206376] env[62521]: _type = "Task" [ 1105.206376] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.213820] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319202, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.345516] env[62521]: DEBUG nova.objects.instance [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'pci_requests' on Instance uuid 7cb9fe57-6940-4f84-bdde-32f185e28451 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.379150] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529ceacd-9124-2de2-6bbb-fd0a4dc54efc, 'name': SearchDatastore_Task, 'duration_secs': 0.008236} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.379435] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.379693] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] e78762d1-a443-443e-ae8c-ab7aac7874bb/e78762d1-a443-443e-ae8c-ab7aac7874bb.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1105.380011] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e225f6b3-cb73-4d78-ae07-6afba616c08b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.386676] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1105.386676] env[62521]: value = "task-1319206" [ 1105.386676] env[62521]: _type = "Task" [ 1105.386676] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.393774] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319206, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.467406] env[62521]: DEBUG oslo_concurrency.lockutils [req-cf7ccfad-4a4e-4eec-9795-96a076bef74d req-282da598-6e29-416b-98dc-16eedf46f538 service nova] Releasing lock "refresh_cache-e78762d1-a443-443e-ae8c-ab7aac7874bb" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.472176] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: d574f077-b221-4f91-8b54-0915421cb36f] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.716247] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319202, 'name': Rename_Task, 'duration_secs': 0.141554} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.716642] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1105.716940] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce66176e-fb3e-4eb5-8893-68967b1720b5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.724506] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1105.724506] env[62521]: value = "task-1319207" [ 1105.724506] env[62521]: _type = "Task" [ 1105.724506] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.734449] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319207, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.851461] env[62521]: DEBUG nova.objects.base [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Object Instance<7cb9fe57-6940-4f84-bdde-32f185e28451> lazy-loaded attributes: flavor,pci_requests {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1105.851749] env[62521]: DEBUG nova.network.neutron [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1105.895309] env[62521]: DEBUG nova.policy [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ce6cad0be5f455eab09b1d348268329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '961924204a584b61a4a85a761821910d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1105.900790] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319206, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.975788] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: acfd78aa-a51e-4a1c-b01a-15854fe0635c] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.201159] env[62521]: DEBUG nova.network.neutron [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Successfully created port: bd7038f6-21a5-429f-bd3b-9ad75c94b49d {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1106.235735] env[62521]: DEBUG oslo_vmware.api [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319207, 'name': PowerOnVM_Task, 'duration_secs': 0.506504} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.236101] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1106.342274] env[62521]: DEBUG nova.compute.manager [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1106.343379] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3674d76-d2ad-4b65-bcf3-e56229b6a917 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.397073] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319206, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524444} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.397360] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] e78762d1-a443-443e-ae8c-ab7aac7874bb/e78762d1-a443-443e-ae8c-ab7aac7874bb.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1106.397570] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1106.397818] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b474b01-bfe8-4bde-8267-ac41785e1f59 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.404753] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1106.404753] env[62521]: value = "task-1319208" [ 1106.404753] env[62521]: _type = "Task" [ 1106.404753] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.411864] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319208, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.479880] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 3805d897-fcb5-43ed-b502-7fad71124e74] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.861396] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26161096-0aac-4cf1-9335-3e9d90999ecf tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.322s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.914886] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319208, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.349956} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.915182] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1106.915952] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20eb6e1-0fe4-4589-a9cb-fd79fdbf482f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.940244] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] e78762d1-a443-443e-ae8c-ab7aac7874bb/e78762d1-a443-443e-ae8c-ab7aac7874bb.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.940532] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-514d85a5-abe3-4ad0-b4a8-c1fa5e959249 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.960404] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1106.960404] env[62521]: value = "task-1319209" [ 1106.960404] env[62521]: _type = "Task" [ 1106.960404] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.967685] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319209, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.983286] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7b5afeab-f04e-453a-a3ed-09e3a725e8d6] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.470438] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319209, 'name': ReconfigVM_Task, 'duration_secs': 0.375446} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.470754] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Reconfigured VM instance instance-0000006b to attach disk [datastore1] e78762d1-a443-443e-ae8c-ab7aac7874bb/e78762d1-a443-443e-ae8c-ab7aac7874bb.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1107.471399] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52cc05e1-70aa-4bbe-bdb0-8396cdaba98a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.477152] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1107.477152] env[62521]: value = "task-1319211" [ 1107.477152] env[62521]: _type = "Task" [ 1107.477152] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.485927] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 2a0ccda4-87a6-4bb3-bf34-4434faf574e6] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.487688] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319211, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.606037] env[62521]: DEBUG nova.compute.manager [req-2840fdf1-3f9b-4468-97fd-6dc4cd3c22c0 req-29cb2cb1-5ff3-4316-84e6-5b8beac8f794 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received event network-vif-plugged-bd7038f6-21a5-429f-bd3b-9ad75c94b49d {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1107.606247] env[62521]: DEBUG oslo_concurrency.lockutils [req-2840fdf1-3f9b-4468-97fd-6dc4cd3c22c0 req-29cb2cb1-5ff3-4316-84e6-5b8beac8f794 service nova] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.606480] env[62521]: DEBUG oslo_concurrency.lockutils [req-2840fdf1-3f9b-4468-97fd-6dc4cd3c22c0 req-29cb2cb1-5ff3-4316-84e6-5b8beac8f794 service nova] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.606668] env[62521]: DEBUG oslo_concurrency.lockutils [req-2840fdf1-3f9b-4468-97fd-6dc4cd3c22c0 req-29cb2cb1-5ff3-4316-84e6-5b8beac8f794 service nova] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.606853] env[62521]: DEBUG nova.compute.manager [req-2840fdf1-3f9b-4468-97fd-6dc4cd3c22c0 req-29cb2cb1-5ff3-4316-84e6-5b8beac8f794 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] No waiting events found dispatching network-vif-plugged-bd7038f6-21a5-429f-bd3b-9ad75c94b49d {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1107.607051] env[62521]: WARNING nova.compute.manager [req-2840fdf1-3f9b-4468-97fd-6dc4cd3c22c0 req-29cb2cb1-5ff3-4316-84e6-5b8beac8f794 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received unexpected event network-vif-plugged-bd7038f6-21a5-429f-bd3b-9ad75c94b49d for instance with vm_state active and task_state None. [ 1107.986974] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319211, 'name': Rename_Task, 'duration_secs': 0.134345} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.988040] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1107.988138] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90f4a9cc-5ed7-4b6b-ba48-4abd79e45e44 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.989729] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 73bc1cf1-7ee1-4a9b-8214-2dc752b700ad] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.996683] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1107.996683] env[62521]: value = "task-1319212" [ 1107.996683] env[62521]: _type = "Task" [ 1107.996683] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.004076] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319212, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.164451] env[62521]: DEBUG nova.network.neutron [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Successfully updated port: bd7038f6-21a5-429f-bd3b-9ad75c94b49d {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1108.186407] env[62521]: DEBUG nova.compute.manager [req-98b57e9e-c4a7-4bf2-9435-3ef60d7f82f7 req-b5b75a5b-d49b-4052-b146-11769ef5ab1d service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received event network-changed-bd7038f6-21a5-429f-bd3b-9ad75c94b49d {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.186562] env[62521]: DEBUG nova.compute.manager [req-98b57e9e-c4a7-4bf2-9435-3ef60d7f82f7 req-b5b75a5b-d49b-4052-b146-11769ef5ab1d service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Refreshing instance network info cache due to event network-changed-bd7038f6-21a5-429f-bd3b-9ad75c94b49d. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1108.186762] env[62521]: DEBUG oslo_concurrency.lockutils [req-98b57e9e-c4a7-4bf2-9435-3ef60d7f82f7 req-b5b75a5b-d49b-4052-b146-11769ef5ab1d service nova] Acquiring lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.186882] env[62521]: DEBUG oslo_concurrency.lockutils [req-98b57e9e-c4a7-4bf2-9435-3ef60d7f82f7 req-b5b75a5b-d49b-4052-b146-11769ef5ab1d service nova] Acquired lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.187078] env[62521]: DEBUG nova.network.neutron [req-98b57e9e-c4a7-4bf2-9435-3ef60d7f82f7 req-b5b75a5b-d49b-4052-b146-11769ef5ab1d service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Refreshing network info cache for port bd7038f6-21a5-429f-bd3b-9ad75c94b49d {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1108.493029] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: bb4afe5a-deef-4c50-a6a8-fbfc3b350fb0] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.506322] env[62521]: DEBUG oslo_vmware.api [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319212, 'name': PowerOnVM_Task, 'duration_secs': 0.440794} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.506604] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1108.506810] env[62521]: INFO nova.compute.manager [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Took 6.71 seconds to spawn the instance on the hypervisor. [ 1108.506988] env[62521]: DEBUG nova.compute.manager [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1108.507782] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08026cf4-1309-4957-a1f9-c0cee416e196 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.666857] env[62521]: DEBUG oslo_concurrency.lockutils [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.878717] env[62521]: DEBUG nova.network.neutron [req-98b57e9e-c4a7-4bf2-9435-3ef60d7f82f7 req-b5b75a5b-d49b-4052-b146-11769ef5ab1d service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Added VIF to instance network info cache for port bd7038f6-21a5-429f-bd3b-9ad75c94b49d. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 1108.879315] env[62521]: DEBUG nova.network.neutron [req-98b57e9e-c4a7-4bf2-9435-3ef60d7f82f7 req-b5b75a5b-d49b-4052-b146-11769ef5ab1d service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bd7038f6-21a5-429f-bd3b-9ad75c94b49d", "address": "fa:16:3e:aa:d4:08", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd7038f6-21", "ovs_interfaceid": "bd7038f6-21a5-429f-bd3b-9ad75c94b49d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.996661] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 92629ebf-eb50-4862-abe2-01e210e302ad] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.024363] env[62521]: INFO nova.compute.manager [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Took 12.56 seconds to build instance. [ 1109.382879] env[62521]: DEBUG oslo_concurrency.lockutils [req-98b57e9e-c4a7-4bf2-9435-3ef60d7f82f7 req-b5b75a5b-d49b-4052-b146-11769ef5ab1d service nova] Releasing lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.383380] env[62521]: DEBUG oslo_concurrency.lockutils [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.383795] env[62521]: DEBUG nova.network.neutron [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1109.500322] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 583b1a1b-1817-45fc-aa4f-fc9de4b4a243] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.525999] env[62521]: DEBUG oslo_concurrency.lockutils [None req-36306825-564a-4f22-9b38-4b4985fa07ce tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "e78762d1-a443-443e-ae8c-ab7aac7874bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.067s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.643308] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1109.643489] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282198', 'volume_id': '0796a138-bf1a-4af3-82d5-800c1f80cba7', 'name': 'volume-0796a138-bf1a-4af3-82d5-800c1f80cba7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '722aa42e-f99c-42a6-aebc-f5ea99447e2f', 'attached_at': '', 'detached_at': '', 'volume_id': '0796a138-bf1a-4af3-82d5-800c1f80cba7', 'serial': '0796a138-bf1a-4af3-82d5-800c1f80cba7'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1109.644458] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050659bd-c9b2-41bf-a0e7-4dbc80499559 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.661338] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7cb179-4bdf-483f-845b-d0ab3e5f47d5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.685635] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] volume-0796a138-bf1a-4af3-82d5-800c1f80cba7/volume-0796a138-bf1a-4af3-82d5-800c1f80cba7.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1109.686302] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-116fc393-af7a-4b8f-a3cc-076dfe7078b8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.704557] env[62521]: DEBUG oslo_vmware.api [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1109.704557] env[62521]: value = "task-1319213" [ 1109.704557] env[62521]: _type = "Task" [ 1109.704557] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.712915] env[62521]: DEBUG oslo_vmware.api [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319213, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.918471] env[62521]: WARNING nova.network.neutron [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] 67bdad13-8ff4-4bf8-b255-d0a285416fbe already exists in list: networks containing: ['67bdad13-8ff4-4bf8-b255-d0a285416fbe']. ignoring it [ 1109.918618] env[62521]: WARNING nova.network.neutron [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] 67bdad13-8ff4-4bf8-b255-d0a285416fbe already exists in list: networks containing: ['67bdad13-8ff4-4bf8-b255-d0a285416fbe']. ignoring it [ 1109.918826] env[62521]: WARNING nova.network.neutron [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] bd7038f6-21a5-429f-bd3b-9ad75c94b49d already exists in list: port_ids containing: ['bd7038f6-21a5-429f-bd3b-9ad75c94b49d']. ignoring it [ 1110.003801] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 4afe09db-6c01-444f-a127-6e1f97794544] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.214598] env[62521]: DEBUG oslo_vmware.api [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319213, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.217885] env[62521]: DEBUG nova.network.neutron [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bd7038f6-21a5-429f-bd3b-9ad75c94b49d", "address": "fa:16:3e:aa:d4:08", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd7038f6-21", "ovs_interfaceid": "bd7038f6-21a5-429f-bd3b-9ad75c94b49d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.277410] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1decf3-b3db-4269-b4da-87cb3ba34a24 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.284417] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d3e4d0c6-4e41-4e03-b429-fc0762f94f96 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Suspending the VM {{(pid=62521) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1110.284650] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-c0997c97-77f5-4eb7-901a-3c5602720abf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.290776] env[62521]: DEBUG oslo_vmware.api [None req-d3e4d0c6-4e41-4e03-b429-fc0762f94f96 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1110.290776] env[62521]: value = "task-1319214" [ 1110.290776] env[62521]: _type = "Task" [ 1110.290776] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.298031] env[62521]: DEBUG oslo_vmware.api [None req-d3e4d0c6-4e41-4e03-b429-fc0762f94f96 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319214, 'name': SuspendVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.507807] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 70e33ae8-3d2f-4b8d-8a1c-22e6e846b8b6] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.715304] env[62521]: DEBUG oslo_vmware.api [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319213, 'name': ReconfigVM_Task, 'duration_secs': 0.822652} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.715676] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfigured VM instance instance-00000066 to attach disk [datastore1] volume-0796a138-bf1a-4af3-82d5-800c1f80cba7/volume-0796a138-bf1a-4af3-82d5-800c1f80cba7.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.720667] env[62521]: DEBUG oslo_concurrency.lockutils [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.721299] env[62521]: DEBUG oslo_concurrency.lockutils [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.721458] env[62521]: DEBUG oslo_concurrency.lockutils [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.721691] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e8ac59c-8173-4839-beb7-c3457c9c942c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.733360] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0631badd-9f47-4594-96b4-39684e330fdf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.751540] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1110.751796] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1110.751964] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1110.752190] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1110.752364] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1110.752531] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1110.752743] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1110.752910] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1110.753123] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1110.753314] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1110.753512] env[62521]: DEBUG nova.virt.hardware [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1110.759913] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Reconfiguring VM to attach interface {{(pid=62521) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1110.761504] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d2d4b03-a519-4c38-a757-996a736ba5e4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.773894] env[62521]: DEBUG oslo_vmware.api [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1110.773894] env[62521]: value = "task-1319215" [ 1110.773894] env[62521]: _type = "Task" [ 1110.773894] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.778781] env[62521]: DEBUG oslo_vmware.api [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1110.778781] env[62521]: value = "task-1319216" [ 1110.778781] env[62521]: _type = "Task" [ 1110.778781] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.785740] env[62521]: DEBUG oslo_vmware.api [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319215, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.790301] env[62521]: DEBUG oslo_vmware.api [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319216, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.799579] env[62521]: DEBUG oslo_vmware.api [None req-d3e4d0c6-4e41-4e03-b429-fc0762f94f96 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319214, 'name': SuspendVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.011979] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 4dcd1c6c-8726-42db-997a-e78ba1293310] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.033768] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.034100] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.285704] env[62521]: DEBUG oslo_vmware.api [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319215, 'name': ReconfigVM_Task, 'duration_secs': 0.173394} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.286319] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282198', 'volume_id': '0796a138-bf1a-4af3-82d5-800c1f80cba7', 'name': 'volume-0796a138-bf1a-4af3-82d5-800c1f80cba7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '722aa42e-f99c-42a6-aebc-f5ea99447e2f', 'attached_at': '', 'detached_at': '', 'volume_id': '0796a138-bf1a-4af3-82d5-800c1f80cba7', 'serial': '0796a138-bf1a-4af3-82d5-800c1f80cba7'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1111.290834] env[62521]: DEBUG oslo_vmware.api [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319216, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.299870] env[62521]: DEBUG oslo_vmware.api [None req-d3e4d0c6-4e41-4e03-b429-fc0762f94f96 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319214, 'name': SuspendVM_Task, 'duration_secs': 0.813551} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.300132] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d3e4d0c6-4e41-4e03-b429-fc0762f94f96 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Suspended the VM {{(pid=62521) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1111.300342] env[62521]: DEBUG nova.compute.manager [None req-d3e4d0c6-4e41-4e03-b429-fc0762f94f96 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1111.301598] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e55ba60-9aa0-4c03-8f6e-48f16353fac3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.515713] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: d5f6492d-ca1b-4e74-b792-b04d55c33660] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.537157] env[62521]: DEBUG nova.compute.utils [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1111.790235] env[62521]: DEBUG oslo_vmware.api [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319216, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.019273] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: af91de53-2ef8-48d1-b8c1-7bd0809f4ae4] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.040591] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.293121] env[62521]: DEBUG oslo_vmware.api [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319216, 'name': ReconfigVM_Task, 'duration_secs': 1.111284} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.293121] env[62521]: DEBUG oslo_concurrency.lockutils [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.293361] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Reconfigured VM to attach interface {{(pid=62521) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1112.325287] env[62521]: DEBUG nova.objects.instance [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'flavor' on Instance uuid 722aa42e-f99c-42a6-aebc-f5ea99447e2f {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.523074] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: b3e33bab-636f-42df-ac6c-667464312c11] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.798177] env[62521]: DEBUG oslo_concurrency.lockutils [None req-84b6d122-b305-433d-bb36-61c80deb7a15 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.028s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.830769] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2508a923-03ef-4f1e-8713-17cae46f9650 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.783s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.026673] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: e8cabcdc-d728-434d-a40d-fe751db086c3] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.105113] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.105415] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.105687] env[62521]: INFO nova.compute.manager [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Attaching volume 7dc3ea6c-838b-490e-a912-2ff188275472 to /dev/sdb [ 1113.135862] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70487fff-0140-48bc-a02c-e0c37fe7c757 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.145490] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb61727-1814-4fd9-9477-aef3df65a4c0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.162759] env[62521]: DEBUG nova.virt.block_device [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Updating existing volume attachment record: 260aaf02-7ace-443e-977b-3c9daf9b29de {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1113.316431] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "e78762d1-a443-443e-ae8c-ab7aac7874bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.316696] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "e78762d1-a443-443e-ae8c-ab7aac7874bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.316907] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "e78762d1-a443-443e-ae8c-ab7aac7874bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.317108] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "e78762d1-a443-443e-ae8c-ab7aac7874bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.317317] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "e78762d1-a443-443e-ae8c-ab7aac7874bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.321501] env[62521]: INFO nova.compute.manager [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Terminating instance [ 1113.323579] env[62521]: DEBUG nova.compute.manager [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1113.323789] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1113.324676] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15358336-c04e-468a-9dbb-e2725c09669f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.333555] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1113.333750] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc8dee9d-cf7d-4f6b-9993-6bf0b1411db6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.410317] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1113.410594] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1113.410794] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleting the datastore file [datastore1] e78762d1-a443-443e-ae8c-ab7aac7874bb {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1113.411087] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf5e76ff-7f5b-47cd-9309-76da29ce19a8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.418438] env[62521]: DEBUG oslo_vmware.api [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1113.418438] env[62521]: value = "task-1319219" [ 1113.418438] env[62521]: _type = "Task" [ 1113.418438] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.427112] env[62521]: DEBUG oslo_vmware.api [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319219, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.530330] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 538ed18e-920e-4ed4-a9e1-a6bfbae7f282] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.930425] env[62521]: DEBUG oslo_vmware.api [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319219, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164524} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.930703] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1113.930894] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1113.931094] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1113.931278] env[62521]: INFO nova.compute.manager [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1113.931567] env[62521]: DEBUG oslo.service.loopingcall [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1113.931788] env[62521]: DEBUG nova.compute.manager [-] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1113.931884] env[62521]: DEBUG nova.network.neutron [-] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1114.034255] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 50aef23e-9429-4499-9fc1-c31b97e23029] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.224579] env[62521]: DEBUG nova.compute.manager [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Stashing vm_state: active {{(pid=62521) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1114.344682] env[62521]: DEBUG nova.compute.manager [req-3202fb86-c0b4-422b-814b-76229a4c8c73 req-184bae34-cc29-47d4-a921-29067901f9db service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Received event network-vif-deleted-13d38661-4ce7-48de-a860-47d294c7548e {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1114.344902] env[62521]: INFO nova.compute.manager [req-3202fb86-c0b4-422b-814b-76229a4c8c73 req-184bae34-cc29-47d4-a921-29067901f9db service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Neutron deleted interface 13d38661-4ce7-48de-a860-47d294c7548e; detaching it from the instance and deleting it from the info cache [ 1114.345092] env[62521]: DEBUG nova.network.neutron [req-3202fb86-c0b4-422b-814b-76229a4c8c73 req-184bae34-cc29-47d4-a921-29067901f9db service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.537121] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 868c7798-8d0f-4c31-88dc-5007454796f1] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.744739] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.744999] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.813615] env[62521]: DEBUG nova.network.neutron [-] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.848111] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ceb3f2ec-c260-4c7b-9185-8cbfb7b3b878 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.859481] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f024f2-3395-44f5-81c1-5e98c6a9eac5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.890074] env[62521]: DEBUG nova.compute.manager [req-3202fb86-c0b4-422b-814b-76229a4c8c73 req-184bae34-cc29-47d4-a921-29067901f9db service nova] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Detach interface failed, port_id=13d38661-4ce7-48de-a860-47d294c7548e, reason: Instance e78762d1-a443-443e-ae8c-ab7aac7874bb could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1114.924938] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-010dd5db-074f-4051-9787-abf0265a5522" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.925220] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-010dd5db-074f-4051-9787-abf0265a5522" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.925588] env[62521]: DEBUG nova.objects.instance [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'flavor' on Instance uuid 7cb9fe57-6940-4f84-bdde-32f185e28451 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.041013] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 062f311c-a783-41cf-8d83-b3a8d4df14f4] Instance has had 0 of 5 cleanup attempts {{(pid=62521) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1115.250118] env[62521]: INFO nova.compute.claims [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1115.316443] env[62521]: INFO nova.compute.manager [-] [instance: e78762d1-a443-443e-ae8c-ab7aac7874bb] Took 1.38 seconds to deallocate network for instance. [ 1115.544851] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1115.545073] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Cleaning up deleted instances with incomplete migration {{(pid=62521) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1115.557932] env[62521]: DEBUG nova.objects.instance [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'pci_requests' on Instance uuid 7cb9fe57-6940-4f84-bdde-32f185e28451 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.757108] env[62521]: INFO nova.compute.resource_tracker [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating resource usage from migration 57897ab4-2b6f-4901-beff-1abddb16fcac [ 1115.823110] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.883628] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd34f6ff-1fdc-4eaa-988b-db1f08abafb3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.891856] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1431a2f4-0d0d-4b22-a7b2-9c3e31b9f0cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.920622] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af4bf73-bd98-4355-a6c4-9b732d622bfb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.927660] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4e25be-f15b-4462-85b1-799f818e46b7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.942135] env[62521]: DEBUG nova.compute.provider_tree [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.047716] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.060541] env[62521]: DEBUG nova.objects.base [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Object Instance<7cb9fe57-6940-4f84-bdde-32f185e28451> lazy-loaded attributes: flavor,pci_requests {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1116.060897] env[62521]: DEBUG nova.network.neutron [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1116.750852] env[62521]: DEBUG nova.scheduler.client.report [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1116.817744] env[62521]: DEBUG nova.policy [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ce6cad0be5f455eab09b1d348268329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '961924204a584b61a4a85a761821910d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1117.255985] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.511s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.256332] env[62521]: INFO nova.compute.manager [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Migrating [ 1117.262981] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.440s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.263283] env[62521]: DEBUG nova.objects.instance [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lazy-loading 'resources' on Instance uuid e78762d1-a443-443e-ae8c-ab7aac7874bb {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.705963] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1117.706244] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282199', 'volume_id': '7dc3ea6c-838b-490e-a912-2ff188275472', 'name': 'volume-7dc3ea6c-838b-490e-a912-2ff188275472', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3aa60da8-b251-4f31-bf8d-a4cd143b491e', 'attached_at': '', 'detached_at': '', 'volume_id': '7dc3ea6c-838b-490e-a912-2ff188275472', 'serial': '7dc3ea6c-838b-490e-a912-2ff188275472'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1117.707164] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd2426c-2a85-4c4f-8da0-5df6804e0be5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.723696] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ab1c42-c257-4e82-8211-e683b5f9cfc5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.747781] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] volume-7dc3ea6c-838b-490e-a912-2ff188275472/volume-7dc3ea6c-838b-490e-a912-2ff188275472.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1117.748081] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0074691d-9776-447a-82f1-19ad62809905 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.760806] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.768771] env[62521]: DEBUG oslo_vmware.api [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1117.768771] env[62521]: value = "task-1319221" [ 1117.768771] env[62521]: _type = "Task" [ 1117.768771] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.774288] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.774511] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.774635] env[62521]: DEBUG nova.network.neutron [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1117.783032] env[62521]: DEBUG oslo_vmware.api [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319221, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.902961] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207e6093-22d6-4317-8390-6bea23d886fc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.910919] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2aa64e-7c82-493e-9358-5f185e1a6acc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.942076] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57a2a78-36d4-4550-a23d-e9f4f3735277 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.950300] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184b0b62-2f0a-4cee-b280-2dd62bd14930 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.964137] env[62521]: DEBUG nova.compute.provider_tree [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.192308] env[62521]: DEBUG nova.compute.manager [req-3745ebb4-a214-4b35-8347-605aebd1f95c req-e8d8acec-ca2e-4c9a-8ea9-be18360bb12b service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received event network-vif-plugged-010dd5db-074f-4051-9787-abf0265a5522 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1118.192919] env[62521]: DEBUG oslo_concurrency.lockutils [req-3745ebb4-a214-4b35-8347-605aebd1f95c req-e8d8acec-ca2e-4c9a-8ea9-be18360bb12b service nova] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.193113] env[62521]: DEBUG oslo_concurrency.lockutils [req-3745ebb4-a214-4b35-8347-605aebd1f95c req-e8d8acec-ca2e-4c9a-8ea9-be18360bb12b service nova] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.193292] env[62521]: DEBUG oslo_concurrency.lockutils [req-3745ebb4-a214-4b35-8347-605aebd1f95c req-e8d8acec-ca2e-4c9a-8ea9-be18360bb12b service nova] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.193467] env[62521]: DEBUG nova.compute.manager [req-3745ebb4-a214-4b35-8347-605aebd1f95c req-e8d8acec-ca2e-4c9a-8ea9-be18360bb12b service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] No waiting events found dispatching network-vif-plugged-010dd5db-074f-4051-9787-abf0265a5522 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1118.193636] env[62521]: WARNING nova.compute.manager [req-3745ebb4-a214-4b35-8347-605aebd1f95c req-e8d8acec-ca2e-4c9a-8ea9-be18360bb12b service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received unexpected event network-vif-plugged-010dd5db-074f-4051-9787-abf0265a5522 for instance with vm_state active and task_state None. [ 1118.267115] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.267425] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1118.267425] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Rebuilding the list of instances to heal {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1118.281116] env[62521]: DEBUG oslo_vmware.api [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319221, 'name': ReconfigVM_Task, 'duration_secs': 0.353188} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.281416] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Reconfigured VM instance instance-00000067 to attach disk [datastore1] volume-7dc3ea6c-838b-490e-a912-2ff188275472/volume-7dc3ea6c-838b-490e-a912-2ff188275472.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.287686] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f8901dc-3d3c-480d-b234-33c32a651192 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.298125] env[62521]: DEBUG nova.network.neutron [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Successfully updated port: 010dd5db-074f-4051-9787-abf0265a5522 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1118.306164] env[62521]: DEBUG oslo_vmware.api [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1118.306164] env[62521]: value = "task-1319222" [ 1118.306164] env[62521]: _type = "Task" [ 1118.306164] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.315466] env[62521]: DEBUG oslo_vmware.api [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319222, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.467315] env[62521]: DEBUG nova.scheduler.client.report [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.501856] env[62521]: DEBUG nova.network.neutron [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance_info_cache with network_info: [{"id": "c6383189-f73f-4f87-8424-e8956b34f94c", "address": "fa:16:3e:ba:1a:11", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6383189-f7", "ovs_interfaceid": "c6383189-f73f-4f87-8424-e8956b34f94c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.798932] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.799101] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquired lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.799297] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Forcefully refreshing network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1118.799466] env[62521]: DEBUG nova.objects.instance [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lazy-loading 'info_cache' on Instance uuid 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.801257] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.801349] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.801522] env[62521]: DEBUG nova.network.neutron [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1118.816930] env[62521]: DEBUG oslo_vmware.api [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319222, 'name': ReconfigVM_Task, 'duration_secs': 0.137637} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.817271] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282199', 'volume_id': '7dc3ea6c-838b-490e-a912-2ff188275472', 'name': 'volume-7dc3ea6c-838b-490e-a912-2ff188275472', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3aa60da8-b251-4f31-bf8d-a4cd143b491e', 'attached_at': '', 'detached_at': '', 'volume_id': '7dc3ea6c-838b-490e-a912-2ff188275472', 'serial': '7dc3ea6c-838b-490e-a912-2ff188275472'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1118.971985] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.709s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.004636] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.015026] env[62521]: INFO nova.scheduler.client.report [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted allocations for instance e78762d1-a443-443e-ae8c-ab7aac7874bb [ 1119.350786] env[62521]: WARNING nova.network.neutron [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] 67bdad13-8ff4-4bf8-b255-d0a285416fbe already exists in list: networks containing: ['67bdad13-8ff4-4bf8-b255-d0a285416fbe']. ignoring it [ 1119.351050] env[62521]: WARNING nova.network.neutron [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] 67bdad13-8ff4-4bf8-b255-d0a285416fbe already exists in list: networks containing: ['67bdad13-8ff4-4bf8-b255-d0a285416fbe']. ignoring it [ 1119.522068] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0abb9271-490a-45e7-8a24-53e08e827958 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "e78762d1-a443-443e-ae8c-ab7aac7874bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.205s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.856170] env[62521]: DEBUG nova.objects.instance [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lazy-loading 'flavor' on Instance uuid 3aa60da8-b251-4f31-bf8d-a4cd143b491e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.939408] env[62521]: DEBUG nova.network.neutron [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bd7038f6-21a5-429f-bd3b-9ad75c94b49d", "address": "fa:16:3e:aa:d4:08", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd7038f6-21", "ovs_interfaceid": "bd7038f6-21a5-429f-bd3b-9ad75c94b49d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "010dd5db-074f-4051-9787-abf0265a5522", "address": "fa:16:3e:9d:4a:53", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap010dd5db-07", "ovs_interfaceid": "010dd5db-074f-4051-9787-abf0265a5522", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.222671] env[62521]: DEBUG nova.compute.manager [req-89a7e9c8-1e92-46d0-b564-b2733b37a3b2 req-c1dfe921-3d69-4460-a88e-c3b6cc2d488e service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received event network-changed-010dd5db-074f-4051-9787-abf0265a5522 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1120.222903] env[62521]: DEBUG nova.compute.manager [req-89a7e9c8-1e92-46d0-b564-b2733b37a3b2 req-c1dfe921-3d69-4460-a88e-c3b6cc2d488e service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Refreshing instance network info cache due to event network-changed-010dd5db-074f-4051-9787-abf0265a5522. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1120.223116] env[62521]: DEBUG oslo_concurrency.lockutils [req-89a7e9c8-1e92-46d0-b564-b2733b37a3b2 req-c1dfe921-3d69-4460-a88e-c3b6cc2d488e service nova] Acquiring lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.364187] env[62521]: DEBUG oslo_concurrency.lockutils [None req-47e01822-a7d1-4a78-9c24-dff385dd111e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.259s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.365259] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "9eeeef50-4538-4b45-b14b-46671356adbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.365479] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "9eeeef50-4538-4b45-b14b-46671356adbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.444422] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.445097] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.445266] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.445543] env[62521]: DEBUG oslo_concurrency.lockutils [req-89a7e9c8-1e92-46d0-b564-b2733b37a3b2 req-c1dfe921-3d69-4460-a88e-c3b6cc2d488e service nova] Acquired lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.445742] env[62521]: DEBUG nova.network.neutron [req-89a7e9c8-1e92-46d0-b564-b2733b37a3b2 req-c1dfe921-3d69-4460-a88e-c3b6cc2d488e service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Refreshing network info cache for port 010dd5db-074f-4051-9787-abf0265a5522 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1120.447530] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda2d0e1-b309-4fe7-9303-6ccea7a674b7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.470379] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1120.470620] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1120.470785] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1120.470970] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1120.471137] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1120.471291] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1120.471511] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1120.471695] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1120.471880] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1120.472060] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1120.472242] env[62521]: DEBUG nova.virt.hardware [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1120.478755] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Reconfiguring VM to attach interface {{(pid=62521) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1120.479310] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63dbe6b9-1d14-4fea-8e95-5c40a76e84f4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.503902] env[62521]: DEBUG oslo_vmware.api [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1120.503902] env[62521]: value = "task-1319223" [ 1120.503902] env[62521]: _type = "Task" [ 1120.503902] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.512697] env[62521]: DEBUG oslo_vmware.api [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319223, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.518327] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc78001b-9d1f-4b45-b9cf-ab557aa6f665 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.540697] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance '722aa42e-f99c-42a6-aebc-f5ea99447e2f' progress to 0 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1120.557585] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.557771] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.582525] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating instance_info_cache with network_info: [{"id": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "address": "fa:16:3e:dd:37:93", "network": {"id": "94250445-301b-430c-ad46-0332334df034", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-991216860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ab001785c25414abbaa4bfe3a6a0bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec3f9e71-839a-429d-b211-d3dfc98ca4f6", "external-id": "nsx-vlan-transportzone-5", "segmentation_id": 5, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2f01aec-5d", "ovs_interfaceid": "a2f01aec-5d56-42ef-ac6c-6364f1571174", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.867277] env[62521]: DEBUG nova.compute.manager [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1121.014552] env[62521]: DEBUG oslo_vmware.api [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319223, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.049383] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1121.049715] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3236ad36-4201-49e9-8bdb-12226ce4164c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.060657] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1121.060657] env[62521]: value = "task-1319224" [ 1121.060657] env[62521]: _type = "Task" [ 1121.060657] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.061511] env[62521]: INFO nova.compute.manager [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Detaching volume 7dc3ea6c-838b-490e-a912-2ff188275472 [ 1121.074501] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319224, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.084969] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Releasing lock "refresh_cache-6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.085211] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updated the network info_cache for instance {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1121.085411] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.085594] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.085758] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.085942] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.086094] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.105656] env[62521]: INFO nova.virt.block_device [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Attempting to driver detach volume 7dc3ea6c-838b-490e-a912-2ff188275472 from mountpoint /dev/sdb [ 1121.105963] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1121.106234] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282199', 'volume_id': '7dc3ea6c-838b-490e-a912-2ff188275472', 'name': 'volume-7dc3ea6c-838b-490e-a912-2ff188275472', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3aa60da8-b251-4f31-bf8d-a4cd143b491e', 'attached_at': '', 'detached_at': '', 'volume_id': '7dc3ea6c-838b-490e-a912-2ff188275472', 'serial': '7dc3ea6c-838b-490e-a912-2ff188275472'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1121.107428] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8601f3e-4c01-4c00-8fd6-5700d9d07e82 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.132606] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af290df0-9891-4c5f-9352-c51bc1b45a87 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.141070] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6c232b-b826-45f5-b7cf-511ff9192531 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.173680] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90dc73b8-cb27-4a58-8d0f-df8049b184f6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.192633] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] The volume has not been displaced from its original location: [datastore1] volume-7dc3ea6c-838b-490e-a912-2ff188275472/volume-7dc3ea6c-838b-490e-a912-2ff188275472.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1121.198205] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Reconfiguring VM instance instance-00000067 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1121.198618] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca9cd2cd-cc59-4714-a382-449a3890158c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.215640] env[62521]: DEBUG nova.network.neutron [req-89a7e9c8-1e92-46d0-b564-b2733b37a3b2 req-c1dfe921-3d69-4460-a88e-c3b6cc2d488e service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updated VIF entry in instance network info cache for port 010dd5db-074f-4051-9787-abf0265a5522. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1121.216161] env[62521]: DEBUG nova.network.neutron [req-89a7e9c8-1e92-46d0-b564-b2733b37a3b2 req-c1dfe921-3d69-4460-a88e-c3b6cc2d488e service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bd7038f6-21a5-429f-bd3b-9ad75c94b49d", "address": "fa:16:3e:aa:d4:08", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd7038f6-21", "ovs_interfaceid": "bd7038f6-21a5-429f-bd3b-9ad75c94b49d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "010dd5db-074f-4051-9787-abf0265a5522", "address": "fa:16:3e:9d:4a:53", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap010dd5db-07", "ovs_interfaceid": "010dd5db-074f-4051-9787-abf0265a5522", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.220111] env[62521]: DEBUG oslo_vmware.api [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1121.220111] env[62521]: value = "task-1319225" [ 1121.220111] env[62521]: _type = "Task" [ 1121.220111] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.229718] env[62521]: DEBUG oslo_vmware.api [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319225, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.393628] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.393943] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.395935] env[62521]: INFO nova.compute.claims [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1121.515020] env[62521]: DEBUG oslo_vmware.api [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319223, 'name': ReconfigVM_Task, 'duration_secs': 0.871247} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.515572] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.515797] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Reconfigured VM to attach interface {{(pid=62521) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1121.574611] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319224, 'name': PowerOffVM_Task, 'duration_secs': 0.321994} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.575033] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1121.575270] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance '722aa42e-f99c-42a6-aebc-f5ea99447e2f' progress to 17 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1121.593111] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.719518] env[62521]: DEBUG oslo_concurrency.lockutils [req-89a7e9c8-1e92-46d0-b564-b2733b37a3b2 req-c1dfe921-3d69-4460-a88e-c3b6cc2d488e service nova] Releasing lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.730274] env[62521]: DEBUG oslo_vmware.api [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319225, 'name': ReconfigVM_Task, 'duration_secs': 0.283121} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.730524] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Reconfigured VM instance instance-00000067 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1121.735052] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6480236a-6238-4131-845d-2928d0b0c1e6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.750762] env[62521]: DEBUG oslo_vmware.api [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1121.750762] env[62521]: value = "task-1319226" [ 1121.750762] env[62521]: _type = "Task" [ 1121.750762] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.760166] env[62521]: DEBUG oslo_vmware.api [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319226, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.023463] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8ffefff6-4df4-4714-9bef-2e74346940b0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-010dd5db-074f-4051-9787-abf0265a5522" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.098s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.082229] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1122.082504] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1122.082667] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.082856] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1122.083015] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.083177] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1122.083428] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1122.083571] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1122.083722] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1122.083866] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1122.084058] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.088969] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1f8168a-6949-4606-b179-6e6fa6782664 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.107753] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1122.107753] env[62521]: value = "task-1319227" [ 1122.107753] env[62521]: _type = "Task" [ 1122.107753] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.116266] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319227, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.261643] env[62521]: DEBUG oslo_vmware.api [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319226, 'name': ReconfigVM_Task, 'duration_secs': 0.136917} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.261967] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282199', 'volume_id': '7dc3ea6c-838b-490e-a912-2ff188275472', 'name': 'volume-7dc3ea6c-838b-490e-a912-2ff188275472', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3aa60da8-b251-4f31-bf8d-a4cd143b491e', 'attached_at': '', 'detached_at': '', 'volume_id': '7dc3ea6c-838b-490e-a912-2ff188275472', 'serial': '7dc3ea6c-838b-490e-a912-2ff188275472'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1122.519028] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f40e0d4-d576-4ee7-8acf-41d088623ecb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.528229] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9eae0a7-7e0a-4680-96f0-9b2e400d641c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.558248] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067f886e-5d01-40a1-8578-314a07953952 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.565543] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a987adfa-2c17-405e-bd5c-36566b031e8d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.578498] env[62521]: DEBUG nova.compute.provider_tree [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.617444] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319227, 'name': ReconfigVM_Task, 'duration_secs': 0.147049} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.618351] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance '722aa42e-f99c-42a6-aebc-f5ea99447e2f' progress to 33 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1122.814571] env[62521]: DEBUG nova.objects.instance [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lazy-loading 'flavor' on Instance uuid 3aa60da8-b251-4f31-bf8d-a4cd143b491e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1123.082263] env[62521]: DEBUG nova.scheduler.client.report [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1123.124324] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1123.124547] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1123.124726] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.124916] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1123.125082] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.125236] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1123.125446] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1123.125603] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1123.125770] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1123.125957] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1123.126116] env[62521]: DEBUG nova.virt.hardware [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.132132] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1123.132787] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6cae2279-c632-4c35-b234-670e7950fb1f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.155664] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1123.155664] env[62521]: value = "task-1319228" [ 1123.155664] env[62521]: _type = "Task" [ 1123.155664] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.164582] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319228, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.587055] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.193s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.587464] env[62521]: DEBUG nova.compute.manager [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1123.590051] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.997s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.590236] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.590436] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1123.591445] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1217f65-a547-4241-b8e4-a865ab69f21c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.600503] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b609b473-70a6-43a8-a655-d6c3696cc688 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.614549] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5706cb09-15e6-439a-b17a-5cf907de742f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.621635] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcee818-ea40-414d-92aa-e349113e4c9c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.651975] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180042MB free_disk=133GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1123.652138] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.652334] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.663750] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319228, 'name': ReconfigVM_Task, 'duration_secs': 0.193496} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.663995] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1123.664721] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2ab248-84ba-4f58-b105-129f60c4b8d2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.689386] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 722aa42e-f99c-42a6-aebc-f5ea99447e2f/722aa42e-f99c-42a6-aebc-f5ea99447e2f.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1123.689632] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-472191d2-eccd-4e83-a19a-aad684da2bfe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.708144] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1123.708144] env[62521]: value = "task-1319229" [ 1123.708144] env[62521]: _type = "Task" [ 1123.708144] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.716342] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319229, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.779102] env[62521]: DEBUG oslo_concurrency.lockutils [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-bd7038f6-21a5-429f-bd3b-9ad75c94b49d" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.779397] env[62521]: DEBUG oslo_concurrency.lockutils [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-bd7038f6-21a5-429f-bd3b-9ad75c94b49d" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.824305] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0511d36c-12de-4e6f-9260-50ce76eb3a19 tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.266s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.093041] env[62521]: DEBUG nova.compute.utils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1124.094748] env[62521]: DEBUG nova.compute.manager [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1124.094917] env[62521]: DEBUG nova.network.neutron [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1124.143149] env[62521]: DEBUG nova.policy [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3566601952144168b5d400c41026c81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0dfb5f5bd4646ec95b7c6a2d0434175', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1124.218535] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319229, 'name': ReconfigVM_Task, 'duration_secs': 0.296376} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.218823] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 722aa42e-f99c-42a6-aebc-f5ea99447e2f/722aa42e-f99c-42a6-aebc-f5ea99447e2f.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1124.219123] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance '722aa42e-f99c-42a6-aebc-f5ea99447e2f' progress to 50 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1124.282595] env[62521]: DEBUG oslo_concurrency.lockutils [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.282831] env[62521]: DEBUG oslo_concurrency.lockutils [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.283891] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66327fab-6abb-4e52-97c9-0133b42f5141 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.303017] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632b6561-85ab-4e66-9587-4195619b27ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.332838] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Reconfiguring VM to detach interface {{(pid=62521) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1124.333138] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8423c766-a770-4554-b084-16f065a31fe7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.353670] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1124.353670] env[62521]: value = "task-1319230" [ 1124.353670] env[62521]: _type = "Task" [ 1124.353670] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.364231] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.525712] env[62521]: DEBUG nova.network.neutron [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Successfully created port: 74195062-c22d-4315-8ff8-47c98a866c60 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1124.598486] env[62521]: DEBUG nova.compute.manager [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1124.661737] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Applying migration context for instance 722aa42e-f99c-42a6-aebc-f5ea99447e2f as it has an incoming, in-progress migration 57897ab4-2b6f-4901-beff-1abddb16fcac. Migration status is migrating {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1124.662990] env[62521]: INFO nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating resource usage from migration 57897ab4-2b6f-4901-beff-1abddb16fcac [ 1124.706506] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 703ff423-98b1-4942-b6fd-8b95fe57bd0a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.706656] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 4021ce0d-69ee-4c57-bb18-fd77f339ed72 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.706847] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 7cb9fe57-6940-4f84-bdde-32f185e28451 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.707035] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 3aa60da8-b251-4f31-bf8d-a4cd143b491e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.707166] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance e159807b-c7b0-4d2e-a7be-426ecaf01785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.707302] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.707421] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 1bea570f-f2b2-4e9a-8dbb-418025402d3d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.707532] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Migration 57897ab4-2b6f-4901-beff-1abddb16fcac is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1124.707643] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 722aa42e-f99c-42a6-aebc-f5ea99447e2f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.707753] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 9eeeef50-4538-4b45-b14b-46671356adbf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.707953] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1124.708125] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1124.725220] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189f5bf3-36c6-49c5-ba16-2ebb125a868b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.754082] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e551c948-4f70-4992-bb1b-af059c76fd48 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.775247] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance '722aa42e-f99c-42a6-aebc-f5ea99447e2f' progress to 67 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1124.866411] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.872800] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3dfcb8f-f93f-4eab-9fd2-8090ac36026a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.880061] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d627cce-be7a-46eb-954c-c54370b92d94 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.911458] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a10d7a-d51b-45b0-9705-3d52f3002ca8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.919251] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a8c88f-b71b-4f07-a309-22f6ebd16909 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.932968] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1125.004266] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.004554] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.004771] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.012280] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.012280] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.012280] env[62521]: INFO nova.compute.manager [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Terminating instance [ 1125.012280] env[62521]: DEBUG nova.compute.manager [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1125.012280] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1125.012280] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a4eb09-391a-44df-a412-76f6d4537c3f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.018634] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1125.018875] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c661833-0c22-40cc-aa3a-36b4b3089614 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.026066] env[62521]: DEBUG oslo_vmware.api [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1125.026066] env[62521]: value = "task-1319231" [ 1125.026066] env[62521]: _type = "Task" [ 1125.026066] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.034941] env[62521]: DEBUG oslo_vmware.api [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319231, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.368116] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.455831] env[62521]: ERROR nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [req-232be109-c792-4a06-b3eb-8504ef4518ac] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-232be109-c792-4a06-b3eb-8504ef4518ac"}]} [ 1125.475333] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1125.494964] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1125.495226] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1125.506659] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1125.524891] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1125.536429] env[62521]: DEBUG oslo_vmware.api [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319231, 'name': PowerOffVM_Task, 'duration_secs': 0.157986} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.538980] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1125.539240] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1125.545357] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52d4a586-7c31-4a0b-991a-37bb29c3f1f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.611411] env[62521]: DEBUG nova.compute.manager [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1125.622521] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1125.622882] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1125.623129] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleting the datastore file [datastore1] 3aa60da8-b251-4f31-bf8d-a4cd143b491e {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1125.626117] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-732c04b7-3144-4355-9388-8ca0cd6fd8c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.634286] env[62521]: DEBUG oslo_vmware.api [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for the task: (returnval){ [ 1125.634286] env[62521]: value = "task-1319233" [ 1125.634286] env[62521]: _type = "Task" [ 1125.634286] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.640634] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1125.640894] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1125.641073] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1125.641267] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1125.641419] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1125.641573] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1125.641781] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1125.641963] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1125.642155] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1125.642333] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1125.642506] env[62521]: DEBUG nova.virt.hardware [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1125.643285] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28541497-578b-42cc-8fb8-419125a9ef19 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.650521] env[62521]: DEBUG oslo_vmware.api [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319233, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.655646] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac85650e-1012-4381-9b2e-a1b8a44bd5ee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.690160] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f2e78a-b57b-4d6b-aa15-4b9ba4f522ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.696769] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffac1766-eca9-4506-ae11-94a4a79f3ce5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.725841] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ffeeda-cd20-4026-9c6e-ce7ea1e6bbb0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.733338] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30471d7d-48e5-4386-a456-87959746488a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.747782] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1125.868949] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.976637] env[62521]: DEBUG nova.compute.manager [req-8bf7cd2a-5aed-4226-be89-a2871450f6a7 req-c8cb6e46-5475-4cc9-9d7e-7dcf741a66a5 service nova] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Received event network-vif-plugged-74195062-c22d-4315-8ff8-47c98a866c60 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1125.976857] env[62521]: DEBUG oslo_concurrency.lockutils [req-8bf7cd2a-5aed-4226-be89-a2871450f6a7 req-c8cb6e46-5475-4cc9-9d7e-7dcf741a66a5 service nova] Acquiring lock "9eeeef50-4538-4b45-b14b-46671356adbf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.977107] env[62521]: DEBUG oslo_concurrency.lockutils [req-8bf7cd2a-5aed-4226-be89-a2871450f6a7 req-c8cb6e46-5475-4cc9-9d7e-7dcf741a66a5 service nova] Lock "9eeeef50-4538-4b45-b14b-46671356adbf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.977215] env[62521]: DEBUG oslo_concurrency.lockutils [req-8bf7cd2a-5aed-4226-be89-a2871450f6a7 req-c8cb6e46-5475-4cc9-9d7e-7dcf741a66a5 service nova] Lock "9eeeef50-4538-4b45-b14b-46671356adbf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.977435] env[62521]: DEBUG nova.compute.manager [req-8bf7cd2a-5aed-4226-be89-a2871450f6a7 req-c8cb6e46-5475-4cc9-9d7e-7dcf741a66a5 service nova] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] No waiting events found dispatching network-vif-plugged-74195062-c22d-4315-8ff8-47c98a866c60 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1125.977618] env[62521]: WARNING nova.compute.manager [req-8bf7cd2a-5aed-4226-be89-a2871450f6a7 req-c8cb6e46-5475-4cc9-9d7e-7dcf741a66a5 service nova] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Received unexpected event network-vif-plugged-74195062-c22d-4315-8ff8-47c98a866c60 for instance with vm_state building and task_state spawning. [ 1126.054526] env[62521]: DEBUG nova.compute.manager [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Stashing vm_state: active {{(pid=62521) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1126.107867] env[62521]: DEBUG nova.network.neutron [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Successfully updated port: 74195062-c22d-4315-8ff8-47c98a866c60 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1126.144983] env[62521]: DEBUG oslo_vmware.api [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319233, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.278506] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 152 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1126.278738] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 152 to 153 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1126.278938] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1126.369835] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.442094] env[62521]: DEBUG nova.network.neutron [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Port c6383189-f73f-4f87-8424-e8956b34f94c binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1126.574432] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.612754] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.613067] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.613067] env[62521]: DEBUG nova.network.neutron [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1126.646036] env[62521]: DEBUG oslo_vmware.api [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319233, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.784408] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1126.784613] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.132s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.784885] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.211s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.870267] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.145435] env[62521]: DEBUG oslo_vmware.api [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Task: {'id': task-1319233, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.464531} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.145435] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1127.145626] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1127.145852] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1127.146111] env[62521]: INFO nova.compute.manager [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Took 2.14 seconds to destroy the instance on the hypervisor. [ 1127.146406] env[62521]: DEBUG oslo.service.loopingcall [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.146643] env[62521]: DEBUG nova.compute.manager [-] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1127.146737] env[62521]: DEBUG nova.network.neutron [-] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1127.165575] env[62521]: DEBUG nova.network.neutron [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1127.290642] env[62521]: INFO nova.compute.claims [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1127.313285] env[62521]: DEBUG nova.network.neutron [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance_info_cache with network_info: [{"id": "74195062-c22d-4315-8ff8-47c98a866c60", "address": "fa:16:3e:29:21:07", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74195062-c2", "ovs_interfaceid": "74195062-c22d-4315-8ff8-47c98a866c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.371514] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.466320] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.467051] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.467051] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.630117] env[62521]: DEBUG nova.compute.manager [req-7cdcb6cc-f81c-460e-9316-7a5444aa8ce9 req-ca26f5c3-1028-4ef4-9f4f-0015e384a9b0 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Received event network-vif-deleted-f4903211-c60f-4fa2-9589-a98babc4b5c5 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1127.630117] env[62521]: INFO nova.compute.manager [req-7cdcb6cc-f81c-460e-9316-7a5444aa8ce9 req-ca26f5c3-1028-4ef4-9f4f-0015e384a9b0 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Neutron deleted interface f4903211-c60f-4fa2-9589-a98babc4b5c5; detaching it from the instance and deleting it from the info cache [ 1127.630117] env[62521]: DEBUG nova.network.neutron [req-7cdcb6cc-f81c-460e-9316-7a5444aa8ce9 req-ca26f5c3-1028-4ef4-9f4f-0015e384a9b0 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.802229] env[62521]: INFO nova.compute.resource_tracker [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating resource usage from migration 140d6d52-68b2-4815-95a8-db67cb6f25b6 [ 1127.815471] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.815980] env[62521]: DEBUG nova.compute.manager [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Instance network_info: |[{"id": "74195062-c22d-4315-8ff8-47c98a866c60", "address": "fa:16:3e:29:21:07", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74195062-c2", "ovs_interfaceid": "74195062-c22d-4315-8ff8-47c98a866c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1127.816470] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:21:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '304be4f7-4e36-4468-9ef4-e457341cef18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '74195062-c22d-4315-8ff8-47c98a866c60', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1127.823949] env[62521]: DEBUG oslo.service.loopingcall [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.826822] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1127.827290] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-554c42fa-3492-4149-a72f-986c8ebc1844 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.851522] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1127.851522] env[62521]: value = "task-1319234" [ 1127.851522] env[62521]: _type = "Task" [ 1127.851522] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.862296] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319234, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.871414] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.965175] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe38af24-c2dd-4093-930a-13cafe3cd982 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.975147] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54f88e9-e030-41b7-a4b1-5a2e1ee98638 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.008667] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624a8028-bd3b-4b2e-a0f8-37aaaf327f25 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.012386] env[62521]: DEBUG nova.compute.manager [req-0f5721a0-b6e6-4e62-8b48-0a6a60e20d3f req-c6d22390-2c1f-41d7-8828-7f82ccb7ea68 service nova] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Received event network-changed-74195062-c22d-4315-8ff8-47c98a866c60 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1128.012487] env[62521]: DEBUG nova.compute.manager [req-0f5721a0-b6e6-4e62-8b48-0a6a60e20d3f req-c6d22390-2c1f-41d7-8828-7f82ccb7ea68 service nova] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Refreshing instance network info cache due to event network-changed-74195062-c22d-4315-8ff8-47c98a866c60. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1128.012640] env[62521]: DEBUG oslo_concurrency.lockutils [req-0f5721a0-b6e6-4e62-8b48-0a6a60e20d3f req-c6d22390-2c1f-41d7-8828-7f82ccb7ea68 service nova] Acquiring lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.012784] env[62521]: DEBUG oslo_concurrency.lockutils [req-0f5721a0-b6e6-4e62-8b48-0a6a60e20d3f req-c6d22390-2c1f-41d7-8828-7f82ccb7ea68 service nova] Acquired lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.012950] env[62521]: DEBUG nova.network.neutron [req-0f5721a0-b6e6-4e62-8b48-0a6a60e20d3f req-c6d22390-2c1f-41d7-8828-7f82ccb7ea68 service nova] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Refreshing network info cache for port 74195062-c22d-4315-8ff8-47c98a866c60 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1128.020884] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d6944b-f769-411f-987f-7786f5f7aa20 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.036939] env[62521]: DEBUG nova.compute.provider_tree [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.107643] env[62521]: DEBUG nova.network.neutron [-] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.132341] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff44ba3b-79c4-4d1c-a0a9-3236c33da820 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.143962] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37fe3092-4068-4c74-8237-14b6d4ca319a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.178071] env[62521]: DEBUG nova.compute.manager [req-7cdcb6cc-f81c-460e-9316-7a5444aa8ce9 req-ca26f5c3-1028-4ef4-9f4f-0015e384a9b0 service nova] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Detach interface failed, port_id=f4903211-c60f-4fa2-9589-a98babc4b5c5, reason: Instance 3aa60da8-b251-4f31-bf8d-a4cd143b491e could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1128.361774] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319234, 'name': CreateVM_Task, 'duration_secs': 0.375109} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.361960] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1128.362701] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.362873] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.363229] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1128.363536] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b27b779e-9ab2-4a2f-be38-a1fec7fa8ba0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.371269] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1128.371269] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52607d8a-0024-2974-99d4-00132ca71ac3" [ 1128.371269] env[62521]: _type = "Task" [ 1128.371269] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.374554] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.382099] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52607d8a-0024-2974-99d4-00132ca71ac3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.500630] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.500851] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.501051] env[62521]: DEBUG nova.network.neutron [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1128.539764] env[62521]: DEBUG nova.scheduler.client.report [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.610629] env[62521]: INFO nova.compute.manager [-] [instance: 3aa60da8-b251-4f31-bf8d-a4cd143b491e] Took 1.46 seconds to deallocate network for instance. [ 1128.716185] env[62521]: DEBUG nova.network.neutron [req-0f5721a0-b6e6-4e62-8b48-0a6a60e20d3f req-c6d22390-2c1f-41d7-8828-7f82ccb7ea68 service nova] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updated VIF entry in instance network info cache for port 74195062-c22d-4315-8ff8-47c98a866c60. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1128.716586] env[62521]: DEBUG nova.network.neutron [req-0f5721a0-b6e6-4e62-8b48-0a6a60e20d3f req-c6d22390-2c1f-41d7-8828-7f82ccb7ea68 service nova] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance_info_cache with network_info: [{"id": "74195062-c22d-4315-8ff8-47c98a866c60", "address": "fa:16:3e:29:21:07", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74195062-c2", "ovs_interfaceid": "74195062-c22d-4315-8ff8-47c98a866c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.874321] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.882365] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52607d8a-0024-2974-99d4-00132ca71ac3, 'name': SearchDatastore_Task, 'duration_secs': 0.010517} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.882613] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.882850] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1128.883097] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.883350] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.883435] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.884045] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-11d2e5ee-3510-4215-98e9-1de43913b46a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.892563] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.892753] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1128.893464] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-183f90d0-c3c9-473d-adde-f0c2943bf049 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.898831] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1128.898831] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ec5f9f-25de-d2c0-da45-2f3fcfec6d30" [ 1128.898831] env[62521]: _type = "Task" [ 1128.898831] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.906053] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ec5f9f-25de-d2c0-da45-2f3fcfec6d30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.044626] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.260s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.044843] env[62521]: INFO nova.compute.manager [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Migrating [ 1129.116571] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.116571] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.116728] env[62521]: DEBUG nova.objects.instance [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lazy-loading 'resources' on Instance uuid 3aa60da8-b251-4f31-bf8d-a4cd143b491e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.204210] env[62521]: DEBUG nova.network.neutron [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance_info_cache with network_info: [{"id": "c6383189-f73f-4f87-8424-e8956b34f94c", "address": "fa:16:3e:ba:1a:11", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6383189-f7", "ovs_interfaceid": "c6383189-f73f-4f87-8424-e8956b34f94c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.219719] env[62521]: DEBUG oslo_concurrency.lockutils [req-0f5721a0-b6e6-4e62-8b48-0a6a60e20d3f req-c6d22390-2c1f-41d7-8828-7f82ccb7ea68 service nova] Releasing lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.374474] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.409640] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ec5f9f-25de-d2c0-da45-2f3fcfec6d30, 'name': SearchDatastore_Task, 'duration_secs': 0.009016} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.410462] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e3abafc-afb1-4730-8fd1-15b3ca9c9fcc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.416170] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1129.416170] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ff1f2d-5d2e-bf6c-be0c-790c8d87f08f" [ 1129.416170] env[62521]: _type = "Task" [ 1129.416170] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.423638] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ff1f2d-5d2e-bf6c-be0c-790c8d87f08f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.558996] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.559225] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.559451] env[62521]: DEBUG nova.network.neutron [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1129.706860] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.739684] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa133393-eafb-45db-a7e6-5f8bfc4ae628 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.749117] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9600bca3-259c-4044-90d8-1c8da76a56bc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.779647] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb855c5-a134-46b4-af6a-e6e82106e2e0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.786977] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b62472-9d1c-49c4-8a97-6906f4ae6a11 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.800675] env[62521]: DEBUG nova.compute.provider_tree [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.875167] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.926896] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52ff1f2d-5d2e-bf6c-be0c-790c8d87f08f, 'name': SearchDatastore_Task, 'duration_secs': 0.010251} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.927237] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.927601] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 9eeeef50-4538-4b45-b14b-46671356adbf/9eeeef50-4538-4b45-b14b-46671356adbf.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1129.927874] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4f48327-3020-4b65-af85-c83817a6d83f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.936858] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1129.936858] env[62521]: value = "task-1319235" [ 1129.936858] env[62521]: _type = "Task" [ 1129.936858] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.945790] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319235, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.217678] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f36650-0342-4bfb-bd3d-96d9f4c2ee8c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.228516] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ea2d97-fc72-40b2-9589-5d6351fbb2ec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.297054] env[62521]: DEBUG nova.network.neutron [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance_info_cache with network_info: [{"id": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "address": "fa:16:3e:20:00:55", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff3e4ad4-29", "ovs_interfaceid": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.303792] env[62521]: DEBUG nova.scheduler.client.report [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1130.376507] env[62521]: DEBUG oslo_vmware.api [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319230, 'name': ReconfigVM_Task, 'duration_secs': 5.770207} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.377234] env[62521]: DEBUG oslo_concurrency.lockutils [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.377388] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Reconfigured VM to detach interface {{(pid=62521) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1130.447174] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319235, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.696893] env[62521]: DEBUG nova.compute.manager [req-5bb85edf-7e54-4ca2-a7aa-be1718e7ad43 req-2aef6db2-0aa8-455d-bf6a-51804a26e587 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received event network-vif-deleted-bd7038f6-21a5-429f-bd3b-9ad75c94b49d {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.697144] env[62521]: INFO nova.compute.manager [req-5bb85edf-7e54-4ca2-a7aa-be1718e7ad43 req-2aef6db2-0aa8-455d-bf6a-51804a26e587 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Neutron deleted interface bd7038f6-21a5-429f-bd3b-9ad75c94b49d; detaching it from the instance and deleting it from the info cache [ 1130.697522] env[62521]: DEBUG nova.network.neutron [req-5bb85edf-7e54-4ca2-a7aa-be1718e7ad43 req-2aef6db2-0aa8-455d-bf6a-51804a26e587 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "010dd5db-074f-4051-9787-abf0265a5522", "address": "fa:16:3e:9d:4a:53", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap010dd5db-07", "ovs_interfaceid": "010dd5db-074f-4051-9787-abf0265a5522", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.800022] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.808806] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.692s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.827272] env[62521]: INFO nova.scheduler.client.report [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Deleted allocations for instance 3aa60da8-b251-4f31-bf8d-a4cd143b491e [ 1130.948580] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319235, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.621221} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.948876] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 9eeeef50-4538-4b45-b14b-46671356adbf/9eeeef50-4538-4b45-b14b-46671356adbf.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1130.949167] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1130.949519] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d3aae86f-ce06-487a-bc8e-9a90601e3600 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.960125] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1130.960125] env[62521]: value = "task-1319236" [ 1130.960125] env[62521]: _type = "Task" [ 1130.960125] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.968560] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319236, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.200804] env[62521]: DEBUG nova.compute.manager [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received event network-vif-deleted-010dd5db-074f-4051-9787-abf0265a5522 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1131.200804] env[62521]: INFO nova.compute.manager [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Neutron deleted interface 010dd5db-074f-4051-9787-abf0265a5522; detaching it from the instance and deleting it from the info cache [ 1131.200804] env[62521]: DEBUG nova.network.neutron [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.202051] env[62521]: DEBUG oslo_concurrency.lockutils [req-5bb85edf-7e54-4ca2-a7aa-be1718e7ad43 req-2aef6db2-0aa8-455d-bf6a-51804a26e587 service nova] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.202195] env[62521]: DEBUG oslo_concurrency.lockutils [req-5bb85edf-7e54-4ca2-a7aa-be1718e7ad43 req-2aef6db2-0aa8-455d-bf6a-51804a26e587 service nova] Acquired lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.203551] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f349563-9f90-4122-a5d0-fbc5da98c8a9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.224629] env[62521]: DEBUG oslo_concurrency.lockutils [req-5bb85edf-7e54-4ca2-a7aa-be1718e7ad43 req-2aef6db2-0aa8-455d-bf6a-51804a26e587 service nova] Releasing lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.224813] env[62521]: WARNING nova.compute.manager [req-5bb85edf-7e54-4ca2-a7aa-be1718e7ad43 req-2aef6db2-0aa8-455d-bf6a-51804a26e587 service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Detach interface failed, port_id=bd7038f6-21a5-429f-bd3b-9ad75c94b49d, reason: No device with interface-id bd7038f6-21a5-429f-bd3b-9ad75c94b49d exists on VM: nova.exception.NotFound: No device with interface-id bd7038f6-21a5-429f-bd3b-9ad75c94b49d exists on VM [ 1131.330909] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7700de2d-d603-4a86-9c19-af60ad27df9c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.336766] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cdd7a007-f61b-4a7c-95be-cff659e7c90e tempest-AttachVolumeNegativeTest-1639732865 tempest-AttachVolumeNegativeTest-1639732865-project-member] Lock "3aa60da8-b251-4f31-bf8d-a4cd143b491e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.332s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.356862] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1889d7ba-eca7-44e9-8b8e-a8d77b29854a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.365341] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance '722aa42e-f99c-42a6-aebc-f5ea99447e2f' progress to 83 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1131.471500] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319236, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.208258} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.471727] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1131.472539] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde6744a-860c-4e87-9832-b8454364ad32 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.495844] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 9eeeef50-4538-4b45-b14b-46671356adbf/9eeeef50-4538-4b45-b14b-46671356adbf.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1131.496544] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a49f21a1-0803-476b-accb-4fcac293df4d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.517732] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1131.517732] env[62521]: value = "task-1319237" [ 1131.517732] env[62521]: _type = "Task" [ 1131.517732] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.527206] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319237, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.678815] env[62521]: DEBUG oslo_concurrency.lockutils [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.679159] env[62521]: DEBUG oslo_concurrency.lockutils [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.679388] env[62521]: DEBUG nova.network.neutron [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1131.703561] env[62521]: DEBUG oslo_concurrency.lockutils [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.703739] env[62521]: DEBUG oslo_concurrency.lockutils [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Acquired lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.704606] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe5a450-7efe-4b50-a6de-312182f12cc4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.724091] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44207589-60bd-47f8-9c92-665fee398b3a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.745993] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.751566] env[62521]: DEBUG nova.virt.vmwareapi.vmops [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Reconfiguring VM to detach interface {{(pid=62521) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1131.751833] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af573d33-3d83-4968-b74d-56c95e332dd3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.770471] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Waiting for the task: (returnval){ [ 1131.770471] env[62521]: value = "task-1319239" [ 1131.770471] env[62521]: _type = "Task" [ 1131.770471] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.778546] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.871815] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1131.872212] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1af78271-3474-4472-a17b-4b0a1a3753f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.879323] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1131.879323] env[62521]: value = "task-1319240" [ 1131.879323] env[62521]: _type = "Task" [ 1131.879323] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.888375] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319240, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.027928] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319237, 'name': ReconfigVM_Task, 'duration_secs': 0.324101} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.028241] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 9eeeef50-4538-4b45-b14b-46671356adbf/9eeeef50-4538-4b45-b14b-46671356adbf.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1132.028866] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db0bf47d-c6e0-416b-911e-c111911267ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.035508] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1132.035508] env[62521]: value = "task-1319241" [ 1132.035508] env[62521]: _type = "Task" [ 1132.035508] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.044316] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319241, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.280689] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.315187] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4231b995-a19e-45fb-963d-9c679f984075 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.334726] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance 'e159807b-c7b0-4d2e-a7be-426ecaf01785' progress to 0 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1132.392293] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319240, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.415176] env[62521]: DEBUG nova.network.neutron [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [{"id": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "address": "fa:16:3e:82:4e:7a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2beae869-60", "ovs_interfaceid": "2beae869-6010-48ba-8e25-6b392dd2ec1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.546063] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319241, 'name': Rename_Task, 'duration_secs': 0.154595} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.546378] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1132.546629] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a58d2e2-49b6-45b6-b1eb-51374e9a6261 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.555227] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1132.555227] env[62521]: value = "task-1319242" [ 1132.555227] env[62521]: _type = "Task" [ 1132.555227] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.564703] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.781431] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.842132] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1132.842132] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4d65f63-a584-4af1-8da2-a39c7db7ff17 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.850755] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1132.850755] env[62521]: value = "task-1319243" [ 1132.850755] env[62521]: _type = "Task" [ 1132.850755] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.859684] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.890528] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319240, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.918483] env[62521]: DEBUG oslo_concurrency.lockutils [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.066907] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319242, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.283282] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.360908] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319243, 'name': PowerOffVM_Task, 'duration_secs': 0.200166} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.361222] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1133.361417] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance 'e159807b-c7b0-4d2e-a7be-426ecaf01785' progress to 17 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1133.390092] env[62521]: DEBUG oslo_vmware.api [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319240, 'name': PowerOnVM_Task, 'duration_secs': 1.450861} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.390349] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1133.390545] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d81e31bd-88cb-49ef-8eb0-12e721775452 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance '722aa42e-f99c-42a6-aebc-f5ea99447e2f' progress to 100 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1133.422819] env[62521]: DEBUG oslo_concurrency.lockutils [None req-86f5c607-8da9-4777-9c58-d974de115628 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-7cb9fe57-6940-4f84-bdde-32f185e28451-bd7038f6-21a5-429f-bd3b-9ad75c94b49d" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.643s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.566669] env[62521]: DEBUG oslo_vmware.api [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319242, 'name': PowerOnVM_Task, 'duration_secs': 0.528853} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.567240] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1133.567458] env[62521]: INFO nova.compute.manager [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Took 7.96 seconds to spawn the instance on the hypervisor. [ 1133.567647] env[62521]: DEBUG nova.compute.manager [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1133.568426] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729d6278-8f0f-46b0-bc92-d162d2c1ea1b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.782307] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.868047] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1133.868220] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1133.868359] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1133.868552] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1133.868708] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1133.868861] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1133.869087] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1133.869259] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1133.869433] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1133.869703] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1133.869901] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1133.875335] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33e5092d-fbc1-4753-9e6f-7840b25da397 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.892769] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1133.892769] env[62521]: value = "task-1319244" [ 1133.892769] env[62521]: _type = "Task" [ 1133.892769] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.905625] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319244, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.084693] env[62521]: INFO nova.compute.manager [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Took 12.71 seconds to build instance. [ 1134.284106] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.402538] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319244, 'name': ReconfigVM_Task, 'duration_secs': 0.18215} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.402905] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance 'e159807b-c7b0-4d2e-a7be-426ecaf01785' progress to 33 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1134.586963] env[62521]: DEBUG oslo_concurrency.lockutils [None req-16924342-1ef8-4fa7-97c4-f8c02df1dd23 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "9eeeef50-4538-4b45-b14b-46671356adbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.221s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.783789] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.909640] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1134.909967] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1134.910095] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1134.910284] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1134.910437] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1134.910595] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1134.910842] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1134.911014] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1134.911186] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1134.911354] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1134.911532] env[62521]: DEBUG nova.virt.hardware [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1134.918345] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1134.918948] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de62584f-195c-4710-a3b1-869404075be5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.948450] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1134.948450] env[62521]: value = "task-1319245" [ 1134.948450] env[62521]: _type = "Task" [ 1134.948450] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.960431] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319245, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.288594] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.462857] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319245, 'name': ReconfigVM_Task, 'duration_secs': 0.180595} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.463198] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1135.463983] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f551b82-bf24-4603-9026-9d2c19cd70cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.486298] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] e159807b-c7b0-4d2e-a7be-426ecaf01785/e159807b-c7b0-4d2e-a7be-426ecaf01785.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.486901] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fe90ef3-e841-43cb-9d00-8e2e36c05a36 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.499844] env[62521]: DEBUG nova.network.neutron [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Port c6383189-f73f-4f87-8424-e8956b34f94c binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1135.500096] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.500255] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.500423] env[62521]: DEBUG nova.network.neutron [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1135.509838] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1135.509838] env[62521]: value = "task-1319247" [ 1135.509838] env[62521]: _type = "Task" [ 1135.509838] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.521322] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319247, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.786916] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.020626] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319247, 'name': ReconfigVM_Task, 'duration_secs': 0.356209} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.020626] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Reconfigured VM instance instance-00000069 to attach disk [datastore1] e159807b-c7b0-4d2e-a7be-426ecaf01785/e159807b-c7b0-4d2e-a7be-426ecaf01785.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1136.021774] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance 'e159807b-c7b0-4d2e-a7be-426ecaf01785' progress to 50 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1136.241297] env[62521]: DEBUG nova.network.neutron [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance_info_cache with network_info: [{"id": "c6383189-f73f-4f87-8424-e8956b34f94c", "address": "fa:16:3e:ba:1a:11", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6383189-f7", "ovs_interfaceid": "c6383189-f73f-4f87-8424-e8956b34f94c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.242562] env[62521]: DEBUG nova.compute.manager [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Stashing vm_state: active {{(pid=62521) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1136.287674] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.528597] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb224d70-a9e7-46f0-b189-efdb5b4eeca1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.548643] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48fbc04-5789-4bc7-90b1-24a9c82a0343 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.565989] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance 'e159807b-c7b0-4d2e-a7be-426ecaf01785' progress to 67 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1136.746159] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.766901] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.767206] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.786986] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.858119] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.858725] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.106613] env[62521]: DEBUG nova.network.neutron [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Port ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1137.254954] env[62521]: DEBUG nova.compute.manager [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62521) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1137.271769] env[62521]: INFO nova.compute.claims [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1137.287337] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.362102] env[62521]: DEBUG nova.compute.utils [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1137.777645] env[62521]: INFO nova.compute.resource_tracker [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating resource usage from migration 90675f29-1342-4dfb-b872-f26c26f1cfe4 [ 1137.789780] env[62521]: DEBUG oslo_vmware.api [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Task: {'id': task-1319239, 'name': ReconfigVM_Task, 'duration_secs': 5.786187} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.790017] env[62521]: DEBUG oslo_concurrency.lockutils [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] Releasing lock "7cb9fe57-6940-4f84-bdde-32f185e28451" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.790242] env[62521]: DEBUG nova.virt.vmwareapi.vmops [req-78ae4f57-8271-457e-92d3-e2af296b0b89 req-7bea1560-b5ad-484d-9e3f-0b0e12a13eda service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Reconfigured VM to detach interface {{(pid=62521) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1137.790745] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.045s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.790975] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.791243] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.791422] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.796369] env[62521]: INFO nova.compute.manager [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Terminating instance [ 1137.798057] env[62521]: DEBUG nova.compute.manager [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1137.799188] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1137.799188] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0409a3d8-bd23-41b3-a8ca-cb3b347424b4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.807306] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1137.807593] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4d2b94b-b5e2-405d-acbb-ad22024e40c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.816168] env[62521]: DEBUG oslo_vmware.api [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1137.816168] env[62521]: value = "task-1319249" [ 1137.816168] env[62521]: _type = "Task" [ 1137.816168] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.826285] env[62521]: DEBUG oslo_vmware.api [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319249, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.865163] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.963301] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9d8e52-4d6d-4c0f-90a4-1b4af64a2c5a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.972991] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36627b9c-8a8b-4351-90ec-4f30dea6e71f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.008826] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525b8bf6-4c97-4885-809f-8a3c712a7be7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.017081] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b40040-0eff-44b1-a00c-843d903ad2ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.030635] env[62521]: DEBUG nova.compute.provider_tree [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1138.129043] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "e159807b-c7b0-4d2e-a7be-426ecaf01785-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.129387] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.129387] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.327131] env[62521]: DEBUG oslo_vmware.api [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319249, 'name': PowerOffVM_Task, 'duration_secs': 0.22229} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.327523] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1138.327714] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1138.327974] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f172e40d-4b42-4b64-872b-d4e3be7bc706 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.363709] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.401226] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1138.401721] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1138.401903] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleting the datastore file [datastore2] 7cb9fe57-6940-4f84-bdde-32f185e28451 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1138.402532] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-236ad4d9-6b55-4b8c-9474-d1539ae70c9b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.410915] env[62521]: DEBUG oslo_vmware.api [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1138.410915] env[62521]: value = "task-1319251" [ 1138.410915] env[62521]: _type = "Task" [ 1138.410915] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.419410] env[62521]: DEBUG oslo_vmware.api [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319251, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.534488] env[62521]: DEBUG nova.scheduler.client.report [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1138.921644] env[62521]: DEBUG oslo_vmware.api [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319251, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193224} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.921865] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.922067] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1138.922256] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1138.922437] env[62521]: INFO nova.compute.manager [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1138.922676] env[62521]: DEBUG oslo.service.loopingcall [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.922871] env[62521]: DEBUG nova.compute.manager [-] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1138.922964] env[62521]: DEBUG nova.network.neutron [-] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1139.003242] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.003514] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.003745] env[62521]: INFO nova.compute.manager [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Attaching volume 1c92dcd2-2187-4e77-aac7-70adfbbfd03d to /dev/sdb [ 1139.005760] env[62521]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 010dd5db-074f-4051-9787-abf0265a5522 could not be found.", "detail": ""}} {{(pid=62521) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1139.005986] env[62521]: DEBUG nova.network.neutron [-] Unable to show port 010dd5db-074f-4051-9787-abf0265a5522 as it no longer exists. {{(pid=62521) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1139.034993] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aada5ae-7d87-4616-889e-b1b1bfc85d7d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.039626] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.272s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.039841] env[62521]: INFO nova.compute.manager [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Migrating [ 1139.048088] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.685s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.052166] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc5b440-64fd-45d4-856a-15be4e3549c4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.073667] env[62521]: DEBUG nova.virt.block_device [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updating existing volume attachment record: 602c8ffe-0ce7-4e32-be59-7f7a5c8fb06d {{(pid=62521) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1139.168567] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.168567] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.168567] env[62521]: DEBUG nova.network.neutron [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1139.439604] env[62521]: DEBUG nova.compute.manager [req-5a899b37-98b1-469b-9ae4-8323c27747b6 req-f73d3b4d-9e77-40fe-a838-3c20791d654d service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Received event network-vif-deleted-2beae869-6010-48ba-8e25-6b392dd2ec1f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.439976] env[62521]: INFO nova.compute.manager [req-5a899b37-98b1-469b-9ae4-8323c27747b6 req-f73d3b4d-9e77-40fe-a838-3c20791d654d service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Neutron deleted interface 2beae869-6010-48ba-8e25-6b392dd2ec1f; detaching it from the instance and deleting it from the info cache [ 1139.440209] env[62521]: DEBUG nova.network.neutron [req-5a899b37-98b1-469b-9ae4-8323c27747b6 req-f73d3b4d-9e77-40fe-a838-3c20791d654d service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.548552] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._sync_power_states {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.560031] env[62521]: DEBUG nova.objects.instance [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'migration_context' on Instance uuid 722aa42e-f99c-42a6-aebc-f5ea99447e2f {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.561164] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.562101] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.562101] env[62521]: DEBUG nova.network.neutron [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1139.894046] env[62521]: DEBUG nova.network.neutron [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance_info_cache with network_info: [{"id": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "address": "fa:16:3e:20:00:55", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff3e4ad4-29", "ovs_interfaceid": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.898559] env[62521]: DEBUG nova.network.neutron [-] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.943806] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1c25829-ab98-4a39-a7e6-f0569ed19f83 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.954356] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a970abaa-13f2-4523-bb98-984b5bb29a91 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.988078] env[62521]: DEBUG nova.compute.manager [req-5a899b37-98b1-469b-9ae4-8323c27747b6 req-f73d3b4d-9e77-40fe-a838-3c20791d654d service nova] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Detach interface failed, port_id=2beae869-6010-48ba-8e25-6b392dd2ec1f, reason: Instance 7cb9fe57-6940-4f84-bdde-32f185e28451 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1140.052327] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Getting list of instances from cluster (obj){ [ 1140.052327] env[62521]: value = "domain-c8" [ 1140.052327] env[62521]: _type = "ClusterComputeResource" [ 1140.052327] env[62521]: } {{(pid=62521) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1140.053526] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04f63d5-f60b-4fda-a803-f10e036b2e14 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.075559] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Got total of 7 instances {{(pid=62521) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1140.075716] env[62521]: WARNING nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] While synchronizing instance power states, found 8 instances in the database and 7 instances on the hypervisor. [ 1140.075848] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Triggering sync for uuid 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 {{(pid=62521) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.076045] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Triggering sync for uuid 703ff423-98b1-4942-b6fd-8b95fe57bd0a {{(pid=62521) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.076204] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Triggering sync for uuid 4021ce0d-69ee-4c57-bb18-fd77f339ed72 {{(pid=62521) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.076353] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Triggering sync for uuid 7cb9fe57-6940-4f84-bdde-32f185e28451 {{(pid=62521) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.076502] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Triggering sync for uuid 722aa42e-f99c-42a6-aebc-f5ea99447e2f {{(pid=62521) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.076646] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Triggering sync for uuid e159807b-c7b0-4d2e-a7be-426ecaf01785 {{(pid=62521) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.076822] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Triggering sync for uuid 1bea570f-f2b2-4e9a-8dbb-418025402d3d {{(pid=62521) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.076998] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Triggering sync for uuid 9eeeef50-4538-4b45-b14b-46671356adbf {{(pid=62521) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.077596] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.077816] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.078110] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.078364] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.078530] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.078710] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.078936] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "7cb9fe57-6940-4f84-bdde-32f185e28451" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.079159] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.079338] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.079502] env[62521]: INFO nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] During sync_power_state the instance has a pending task (resize_reverting). Skip. [ 1140.079700] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.079919] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "e159807b-c7b0-4d2e-a7be-426ecaf01785" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.080121] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.080282] env[62521]: INFO nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] During sync_power_state the instance has a pending task (resize_migrated). Skip. [ 1140.080438] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.080625] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.080825] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "9eeeef50-4538-4b45-b14b-46671356adbf" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.081008] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "9eeeef50-4538-4b45-b14b-46671356adbf" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.081177] env[62521]: INFO nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] During sync_power_state the instance has a pending task (resize_prep). Skip. [ 1140.081332] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "9eeeef50-4538-4b45-b14b-46671356adbf" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.082433] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3d7ce6-3a46-41c7-b5f7-4cfbc44431a7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.085307] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e72148a-0ffc-42c9-b4c0-4b54e7cafb7a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.088435] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3e4bf7-a413-48d0-ba26-cf8c23e4cbb2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.234519] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4636595-d5da-4f63-a609-fd92d4e0efbf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.242701] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c84eab-0765-4e76-9d62-22ae275e30af {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.275708] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5609bb-44a6-44fd-bf15-e00d5de51b92 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.283874] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263da103-d0f4-4ef9-b3c7-8cd41e7fd6ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.299406] env[62521]: DEBUG nova.compute.provider_tree [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1140.397425] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.401644] env[62521]: INFO nova.compute.manager [-] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] Took 1.48 seconds to deallocate network for instance. [ 1140.496962] env[62521]: DEBUG nova.network.neutron [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance_info_cache with network_info: [{"id": "74195062-c22d-4315-8ff8-47c98a866c60", "address": "fa:16:3e:29:21:07", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74195062-c2", "ovs_interfaceid": "74195062-c22d-4315-8ff8-47c98a866c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.605108] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.527s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.605566] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.528s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.605871] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.527s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.803331] env[62521]: DEBUG nova.scheduler.client.report [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1140.910354] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.926127] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0154bae-1317-435f-b1d3-ed81f74412ff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.947941] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c11a33-bc77-4719-8745-48f51eeb1261 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.956434] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance 'e159807b-c7b0-4d2e-a7be-426ecaf01785' progress to 83 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1141.001068] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.463742] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1141.464394] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86bb1209-0ff6-4c88-9a53-128876f23d31 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.473280] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1141.473280] env[62521]: value = "task-1319255" [ 1141.473280] env[62521]: _type = "Task" [ 1141.473280] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.484562] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319255, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.814967] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.767s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.822555] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.912s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.823385] env[62521]: DEBUG nova.objects.instance [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'resources' on Instance uuid 7cb9fe57-6940-4f84-bdde-32f185e28451 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.985988] env[62521]: DEBUG oslo_vmware.api [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319255, 'name': PowerOnVM_Task, 'duration_secs': 0.448085} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.986246] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1141.986439] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f23bedf3-fe6d-4f3c-8ea0-8bd4f4490cdd tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance 'e159807b-c7b0-4d2e-a7be-426ecaf01785' progress to 100 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1142.477725] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8e37b1-6272-4c29-affc-c554a214f607 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.485425] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa1df06-fbeb-4e5c-aed8-5a3e1e04f89d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.521742] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045416e9-4e3f-4465-a501-02ca146f11fe {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.525231] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73992024-08a6-4df8-afe0-dca6f1d87076 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.547740] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance '9eeeef50-4538-4b45-b14b-46671356adbf' progress to 0 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1142.554668] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3449b415-c743-4805-ab61-13083cb78b5a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.571113] env[62521]: DEBUG nova.compute.provider_tree [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.055824] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1143.056504] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e62b6c06-2ff6-4602-aa03-b63ba504dfef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.064097] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1143.064097] env[62521]: value = "task-1319257" [ 1143.064097] env[62521]: _type = "Task" [ 1143.064097] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.075522] env[62521]: DEBUG nova.scheduler.client.report [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1143.078928] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319257, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.086234] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.086508] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.086768] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.086977] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.087190] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.089442] env[62521]: INFO nova.compute.manager [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Terminating instance [ 1143.091523] env[62521]: DEBUG nova.compute.manager [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1143.091743] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1143.092621] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c24c067-79ba-4ba3-abe6-4b3aed9c4fb6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.102180] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1143.102342] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2b997a7-fefa-495e-a47b-42a89cb3e883 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.109808] env[62521]: DEBUG oslo_vmware.api [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1143.109808] env[62521]: value = "task-1319258" [ 1143.109808] env[62521]: _type = "Task" [ 1143.109808] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.118256] env[62521]: DEBUG oslo_vmware.api [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319258, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.363035] env[62521]: INFO nova.compute.manager [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Swapping old allocation on dict_keys(['3d21544b-bfc8-42d0-86ca-d323b5e2628f']) held by migration 57897ab4-2b6f-4901-beff-1abddb16fcac for instance [ 1143.387640] env[62521]: DEBUG nova.scheduler.client.report [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Overwriting current allocation {'allocations': {'3d21544b-bfc8-42d0-86ca-d323b5e2628f': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 155}}, 'project_id': '11829634203c4967bbcff90eb904097c', 'user_id': 'dfab7121825d4e049370b526f90eb499', 'consumer_generation': 1} on consumer 722aa42e-f99c-42a6-aebc-f5ea99447e2f {{(pid=62521) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1143.576553] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319257, 'name': PowerOffVM_Task, 'duration_secs': 0.272929} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.576950] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1143.577240] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance '9eeeef50-4538-4b45-b14b-46671356adbf' progress to 17 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1143.581905] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.759s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.593284] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.593470] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.593658] env[62521]: DEBUG nova.network.neutron [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1143.604227] env[62521]: INFO nova.scheduler.client.report [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleted allocations for instance 7cb9fe57-6940-4f84-bdde-32f185e28451 [ 1143.620316] env[62521]: DEBUG oslo_vmware.api [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319258, 'name': PowerOffVM_Task, 'duration_secs': 0.227901} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.621486] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Volume attach. Driver type: vmdk {{(pid=62521) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1143.621705] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282202', 'volume_id': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'name': 'volume-1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1bea570f-f2b2-4e9a-8dbb-418025402d3d', 'attached_at': '', 'detached_at': '', 'volume_id': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'serial': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1143.622024] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1143.622206] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1143.623028] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b33939-1dac-4f8b-b721-9b41fb2deac2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.625455] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1143444a-5ff2-428d-bf0e-40be422dd96f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.644623] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da1dc8e-261d-4b7b-8650-55cdb20ee996 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.671543] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-1c92dcd2-2187-4e77-aac7-70adfbbfd03d/volume-1c92dcd2-2187-4e77-aac7-70adfbbfd03d.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1143.672159] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dab8ed91-26b0-478d-81e8-ca33a9f5fa31 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.691614] env[62521]: DEBUG oslo_vmware.api [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1143.691614] env[62521]: value = "task-1319260" [ 1143.691614] env[62521]: _type = "Task" [ 1143.691614] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.700655] env[62521]: DEBUG oslo_vmware.api [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319260, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.712585] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1143.712874] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1143.713091] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleting the datastore file [datastore2] 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1143.713368] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0be5f2b2-d154-4cdc-9027-77859eb744eb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.723266] env[62521]: DEBUG oslo_vmware.api [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for the task: (returnval){ [ 1143.723266] env[62521]: value = "task-1319261" [ 1143.723266] env[62521]: _type = "Task" [ 1143.723266] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.731967] env[62521]: DEBUG oslo_vmware.api [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319261, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.086760] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1144.087064] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1144.087235] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1144.087427] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1144.087597] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1144.087765] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1144.087978] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1144.088233] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1144.088332] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1144.088499] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1144.088696] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1144.094058] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f4f8496-699e-48fa-932a-1cb67eef587c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.110691] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2f7dc8fc-c129-4d19-bc9c-84c1e5c7a73d tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.320s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.111641] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.033s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.111845] env[62521]: INFO nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 7cb9fe57-6940-4f84-bdde-32f185e28451] During sync_power_state the instance has a pending task (deleting). Skip. [ 1144.112065] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "7cb9fe57-6940-4f84-bdde-32f185e28451" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.114342] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1144.114342] env[62521]: value = "task-1319262" [ 1144.114342] env[62521]: _type = "Task" [ 1144.114342] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.124107] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319262, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.202761] env[62521]: DEBUG oslo_vmware.api [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319260, 'name': ReconfigVM_Task, 'duration_secs': 0.379584} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.202761] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-1c92dcd2-2187-4e77-aac7-70adfbbfd03d/volume-1c92dcd2-2187-4e77-aac7-70adfbbfd03d.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1144.208044] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3226b417-4ac3-410b-a231-1c83a9d9b4f7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.229427] env[62521]: DEBUG oslo_vmware.api [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1144.229427] env[62521]: value = "task-1319263" [ 1144.229427] env[62521]: _type = "Task" [ 1144.229427] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.236707] env[62521]: DEBUG oslo_vmware.api [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Task: {'id': task-1319261, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156489} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.237422] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1144.237652] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1144.237860] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1144.238058] env[62521]: INFO nova.compute.manager [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1144.238324] env[62521]: DEBUG oslo.service.loopingcall [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1144.238536] env[62521]: DEBUG nova.compute.manager [-] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1144.238632] env[62521]: DEBUG nova.network.neutron [-] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1144.244168] env[62521]: DEBUG oslo_vmware.api [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319263, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.392240] env[62521]: DEBUG nova.network.neutron [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance_info_cache with network_info: [{"id": "c6383189-f73f-4f87-8424-e8956b34f94c", "address": "fa:16:3e:ba:1a:11", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6383189-f7", "ovs_interfaceid": "c6383189-f73f-4f87-8424-e8956b34f94c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.628008] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319262, 'name': ReconfigVM_Task, 'duration_secs': 0.200134} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.628381] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance '9eeeef50-4538-4b45-b14b-46671356adbf' progress to 33 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1144.740953] env[62521]: DEBUG oslo_vmware.api [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319263, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.789639] env[62521]: DEBUG nova.compute.manager [req-a2b9f43a-ccc1-41fb-9ed0-106cdbf33857 req-d96e94dd-df6f-479c-ad82-1d6f2b5e312c service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Received event network-vif-deleted-a2f01aec-5d56-42ef-ac6c-6364f1571174 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1144.789883] env[62521]: INFO nova.compute.manager [req-a2b9f43a-ccc1-41fb-9ed0-106cdbf33857 req-d96e94dd-df6f-479c-ad82-1d6f2b5e312c service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Neutron deleted interface a2f01aec-5d56-42ef-ac6c-6364f1571174; detaching it from the instance and deleting it from the info cache [ 1144.790041] env[62521]: DEBUG nova.network.neutron [req-a2b9f43a-ccc1-41fb-9ed0-106cdbf33857 req-d96e94dd-df6f-479c-ad82-1d6f2b5e312c service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.894722] env[62521]: DEBUG oslo_concurrency.lockutils [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-722aa42e-f99c-42a6-aebc-f5ea99447e2f" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.895805] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03557c2c-e729-4572-b994-19b0bac432ba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.903804] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c9cea9-c003-48d0-aeb2-21663633b2cf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.929623] env[62521]: DEBUG nova.network.neutron [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Port ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1144.929623] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.929623] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.929623] env[62521]: DEBUG nova.network.neutron [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1145.135629] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1145.135990] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1145.136197] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1145.136482] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1145.136560] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1145.136711] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1145.136901] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1145.137136] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1145.137329] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1145.137559] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1145.137767] env[62521]: DEBUG nova.virt.hardware [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1145.144506] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1145.144506] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de41a3de-64f8-4593-8341-a6957fa33176 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.169367] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1145.169367] env[62521]: value = "task-1319264" [ 1145.169367] env[62521]: _type = "Task" [ 1145.169367] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.178721] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319264, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.243334] env[62521]: DEBUG oslo_vmware.api [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319263, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.266288] env[62521]: DEBUG nova.network.neutron [-] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.293972] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15dd6b7d-57c5-4d99-885a-8b13a29fb969 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.304570] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7754f18-342c-4512-b1f4-7b549848f832 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.338930] env[62521]: DEBUG nova.compute.manager [req-a2b9f43a-ccc1-41fb-9ed0-106cdbf33857 req-d96e94dd-df6f-479c-ad82-1d6f2b5e312c service nova] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Detach interface failed, port_id=a2f01aec-5d56-42ef-ac6c-6364f1571174, reason: Instance 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1145.674163] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319264, 'name': ReconfigVM_Task, 'duration_secs': 0.243138} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.674477] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1145.675261] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3ae1af-92f2-4079-8b95-75037bb65ea4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.697379] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 9eeeef50-4538-4b45-b14b-46671356adbf/9eeeef50-4538-4b45-b14b-46671356adbf.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1145.699884] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1928a89-d8ab-4b2a-a049-98ed349a0287 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.719700] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1145.719700] env[62521]: value = "task-1319265" [ 1145.719700] env[62521]: _type = "Task" [ 1145.719700] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.728221] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319265, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.739373] env[62521]: DEBUG oslo_vmware.api [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319263, 'name': ReconfigVM_Task, 'duration_secs': 1.167976} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.742062] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282202', 'volume_id': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'name': 'volume-1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1bea570f-f2b2-4e9a-8dbb-418025402d3d', 'attached_at': '', 'detached_at': '', 'volume_id': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'serial': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d'} {{(pid=62521) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1145.769995] env[62521]: INFO nova.compute.manager [-] [instance: 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7] Took 1.53 seconds to deallocate network for instance. [ 1145.816507] env[62521]: DEBUG nova.network.neutron [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance_info_cache with network_info: [{"id": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "address": "fa:16:3e:20:00:55", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff3e4ad4-29", "ovs_interfaceid": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.010020] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1146.010381] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b657386-336c-49fe-bf2f-235b5d74214d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.020258] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1146.020258] env[62521]: value = "task-1319266" [ 1146.020258] env[62521]: _type = "Task" [ 1146.020258] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.029362] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319266, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.231423] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319265, 'name': ReconfigVM_Task, 'duration_secs': 0.354094} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.231779] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 9eeeef50-4538-4b45-b14b-46671356adbf/9eeeef50-4538-4b45-b14b-46671356adbf.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1146.232088] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance '9eeeef50-4538-4b45-b14b-46671356adbf' progress to 50 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1146.277067] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.277795] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.277795] env[62521]: DEBUG nova.objects.instance [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lazy-loading 'resources' on Instance uuid 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.321070] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.530934] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319266, 'name': PowerOffVM_Task, 'duration_secs': 0.198747} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.531368] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1146.532106] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1146.532377] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1146.532588] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1146.532834] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1146.533047] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1146.533251] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1146.533510] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1146.533751] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1146.534019] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1146.534213] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1146.534441] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1146.539399] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6478905b-06b6-4837-ba0e-4349c1749d64 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.556661] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1146.556661] env[62521]: value = "task-1319267" [ 1146.556661] env[62521]: _type = "Task" [ 1146.556661] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.565623] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319267, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.739464] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7d771e-fa12-4f54-8284-a48ad374eff7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.760507] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a38770-c3c0-420d-b308-20c6d555c7d1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.779606] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance '9eeeef50-4538-4b45-b14b-46671356adbf' progress to 67 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1146.786954] env[62521]: DEBUG nova.objects.instance [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lazy-loading 'flavor' on Instance uuid 1bea570f-f2b2-4e9a-8dbb-418025402d3d {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.822804] env[62521]: DEBUG nova.compute.manager [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62521) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1146.823062] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.931477] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e27fc6-94ba-4559-99fe-b55e7f57ba3e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.943702] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6fd5dff-6f1a-4d94-8366-9faf76efafd6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.977009] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.977319] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.980182] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955a199c-e6bc-4d99-955d-f0b6bf2e7380 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.989249] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0daa49df-942f-4426-9197-318df36d9330 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.007369] env[62521]: DEBUG nova.compute.provider_tree [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.066584] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319267, 'name': ReconfigVM_Task, 'duration_secs': 0.258534} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.067711] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894d5a9a-35c7-40e4-946e-f0c3fe4bbeaa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.091249] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1147.092128] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1147.092128] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1147.092128] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1147.092293] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1147.092520] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1147.092783] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1147.092982] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1147.093214] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1147.093423] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1147.093638] env[62521]: DEBUG nova.virt.hardware [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.094530] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d9d3cb2-3f3e-463e-808c-174e81a5d7c0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.101350] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1147.101350] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520d3150-77ed-0575-b650-dd62905e0538" [ 1147.101350] env[62521]: _type = "Task" [ 1147.101350] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.109819] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520d3150-77ed-0575-b650-dd62905e0538, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.294665] env[62521]: DEBUG oslo_concurrency.lockutils [None req-bd859792-c676-41b7-8036-d74e3c645b09 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.291s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.295569] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.215s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.296892] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395dfb75-b901-4e0f-b5d1-7d84ef38cefa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.331318] env[62521]: DEBUG nova.network.neutron [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Port 74195062-c22d-4315-8ff8-47c98a866c60 binding to destination host cpu-1 is already ACTIVE {{(pid=62521) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1147.430018] env[62521]: INFO nova.compute.manager [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Rescuing [ 1147.430342] env[62521]: DEBUG oslo_concurrency.lockutils [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.430502] env[62521]: DEBUG oslo_concurrency.lockutils [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.430676] env[62521]: DEBUG nova.network.neutron [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1147.479747] env[62521]: DEBUG nova.compute.manager [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1147.510576] env[62521]: DEBUG nova.scheduler.client.report [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1147.611715] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]520d3150-77ed-0575-b650-dd62905e0538, 'name': SearchDatastore_Task, 'duration_secs': 0.00754} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.616951] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1147.617234] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf7f3bd9-814f-4b34-8f8f-ed078742c415 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.634990] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1147.634990] env[62521]: value = "task-1319268" [ 1147.634990] env[62521]: _type = "Task" [ 1147.634990] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.642577] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319268, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.808077] env[62521]: INFO nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] During sync_power_state the instance has a pending task (rescuing). Skip. [ 1147.808392] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.513s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.008508] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.015115] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.738s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.017242] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 1.194s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.036176] env[62521]: INFO nova.scheduler.client.report [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Deleted allocations for instance 6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7 [ 1148.145943] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319268, 'name': ReconfigVM_Task, 'duration_secs': 0.2128} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.146372] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1148.147319] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6ec434-4778-483d-8b81-477d7f3c7094 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.175869] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 722aa42e-f99c-42a6-aebc-f5ea99447e2f/722aa42e-f99c-42a6-aebc-f5ea99447e2f.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1148.176510] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dfa74426-8ffb-45d7-92d7-b88785d35b98 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.196620] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1148.196620] env[62521]: value = "task-1319269" [ 1148.196620] env[62521]: _type = "Task" [ 1148.196620] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.206935] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319269, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.227116] env[62521]: DEBUG nova.network.neutron [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updating instance_info_cache with network_info: [{"id": "02aabc96-04f0-4b09-bcf6-44443d6542af", "address": "fa:16:3e:4a:64:63", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aabc96-04", "ovs_interfaceid": "02aabc96-04f0-4b09-bcf6-44443d6542af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.353855] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "9eeeef50-4538-4b45-b14b-46671356adbf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.354124] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "9eeeef50-4538-4b45-b14b-46671356adbf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.354330] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "9eeeef50-4538-4b45-b14b-46671356adbf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.521259] env[62521]: DEBUG nova.objects.instance [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'migration_context' on Instance uuid e159807b-c7b0-4d2e-a7be-426ecaf01785 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1148.542490] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ac72ab2e-61e8-417a-a987-1b8309a7f6b5 tempest-AttachVolumeShelveTestJSON-1369137429 tempest-AttachVolumeShelveTestJSON-1369137429-project-member] Lock "6bd3fd55-76fc-482e-b3c7-5b52e76e9fb7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.456s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.706999] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319269, 'name': ReconfigVM_Task, 'duration_secs': 0.288368} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.707237] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 722aa42e-f99c-42a6-aebc-f5ea99447e2f/722aa42e-f99c-42a6-aebc-f5ea99447e2f.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1148.708114] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73b1d3d-92e4-4567-8920-db4e7c9cafe0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.730288] env[62521]: DEBUG oslo_concurrency.lockutils [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.732800] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a8d287-67fe-4796-911d-95d7614814ab {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.754657] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92c2866-9fc7-4ed9-bb54-abccb3b31268 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.777930] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4227b9-dfa9-4d82-a0d0-c2b27ab84f47 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.784702] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1148.785511] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-692b2dd3-15a3-4c52-b3b7-24abf8f01d46 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.792433] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1148.792433] env[62521]: value = "task-1319270" [ 1148.792433] env[62521]: _type = "Task" [ 1148.792433] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.800232] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319270, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.170655] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb979885-d846-45e0-91f5-274f5a8780df {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.180034] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52149ff-912c-4153-99c9-71f15444da00 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.217591] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776b097b-8b02-4172-9b9b-ec26ff37a9d1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.225968] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50deed2d-80f1-4249-81e4-cbb08d42ec0c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.240885] env[62521]: DEBUG nova.compute.provider_tree [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1149.285434] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1149.285694] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd91bd37-46a7-4581-807d-c5701f96d343 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.294501] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1149.294501] env[62521]: value = "task-1319271" [ 1149.294501] env[62521]: _type = "Task" [ 1149.294501] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.308130] env[62521]: DEBUG oslo_vmware.api [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319270, 'name': PowerOnVM_Task, 'duration_secs': 0.401} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.311754] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1149.316323] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.393172] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.393377] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.393561] env[62521]: DEBUG nova.network.neutron [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1149.744673] env[62521]: DEBUG nova.scheduler.client.report [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1149.808350] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319271, 'name': PowerOffVM_Task, 'duration_secs': 0.321053} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.808561] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1149.809342] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e506c40-e58b-428d-98c2-afc2a5537c32 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.834071] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8496d9d-3d8d-4332-af35-95d65008ebbb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.873230] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1149.873517] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9db46216-8fb5-4dfc-878d-1b117226ee53 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.880804] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1149.880804] env[62521]: value = "task-1319272" [ 1149.880804] env[62521]: _type = "Task" [ 1149.880804] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.888193] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.075169] env[62521]: DEBUG nova.network.neutron [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance_info_cache with network_info: [{"id": "74195062-c22d-4315-8ff8-47c98a866c60", "address": "fa:16:3e:29:21:07", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74195062-c2", "ovs_interfaceid": "74195062-c22d-4315-8ff8-47c98a866c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.376440] env[62521]: INFO nova.compute.manager [None req-b83ab105-58fd-45b6-944f-2f727e00b62d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance to original state: 'active' [ 1150.391648] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1150.391889] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1150.392184] env[62521]: DEBUG oslo_concurrency.lockutils [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.392336] env[62521]: DEBUG oslo_concurrency.lockutils [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.392517] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1150.392773] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-101aca48-6b98-4b5a-b4b7-1a19f1b0c0fb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.402498] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1150.402804] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1150.404501] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21992e8e-a091-448b-be68-5c9b8102a088 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.410670] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1150.410670] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]528d0633-a7bc-8edb-f76e-7914f82636bb" [ 1150.410670] env[62521]: _type = "Task" [ 1150.410670] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.419167] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]528d0633-a7bc-8edb-f76e-7914f82636bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.579635] env[62521]: DEBUG oslo_concurrency.lockutils [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.756088] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.739s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.762599] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.754s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.764177] env[62521]: INFO nova.compute.claims [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1150.923339] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]528d0633-a7bc-8edb-f76e-7914f82636bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009912} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.924155] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37947fae-19aa-4c0b-8f6d-8bc642e3fa0e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.929938] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1150.929938] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526c3df0-75ae-040b-d772-2b5851c4e9da" [ 1150.929938] env[62521]: _type = "Task" [ 1150.929938] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.939062] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526c3df0-75ae-040b-d772-2b5851c4e9da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.125722] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52166a0c-78cb-4daa-8ce2-d21ca2e6bab6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.146942] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18200e9b-70fe-4d2c-948e-aba61c2c5a07 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.155599] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance '9eeeef50-4538-4b45-b14b-46671356adbf' progress to 83 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1151.441203] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]526c3df0-75ae-040b-d772-2b5851c4e9da, 'name': SearchDatastore_Task, 'duration_secs': 0.013835} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.441555] env[62521]: DEBUG oslo_concurrency.lockutils [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.441751] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 1bea570f-f2b2-4e9a-8dbb-418025402d3d/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk. {{(pid=62521) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1151.442034] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-592c0e8a-eeb2-4d14-96a5-60ed3d3a3159 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.450122] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1151.450122] env[62521]: value = "task-1319274" [ 1151.450122] env[62521]: _type = "Task" [ 1151.450122] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.459046] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319274, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.573183] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.573499] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.573735] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.573948] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.574147] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.576648] env[62521]: INFO nova.compute.manager [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Terminating instance [ 1151.579107] env[62521]: DEBUG nova.compute.manager [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1151.579334] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1151.579581] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79273395-62df-40d5-a95c-e25c440ab952 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.586857] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1151.586857] env[62521]: value = "task-1319275" [ 1151.586857] env[62521]: _type = "Task" [ 1151.586857] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.598719] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319275, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.662082] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1151.662395] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc9bd823-d124-4819-a9e0-23f74d7ed5b3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.671122] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1151.671122] env[62521]: value = "task-1319276" [ 1151.671122] env[62521]: _type = "Task" [ 1151.671122] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.680881] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319276, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.939422] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846b207d-064a-4600-9a6c-0b63f4d3a769 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.949043] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5403f4fa-3955-429a-b78b-da7f1f8bda4d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.961310] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319274, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.987565] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41d9566-23df-4eda-b7b4-c99cfc578a65 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.995400] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278e8c4a-ec50-488a-9c94-f8a7f5073657 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.009583] env[62521]: DEBUG nova.compute.provider_tree [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.097719] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319275, 'name': PowerOffVM_Task, 'duration_secs': 0.254522} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.098081] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1152.098358] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1152.098599] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282198', 'volume_id': '0796a138-bf1a-4af3-82d5-800c1f80cba7', 'name': 'volume-0796a138-bf1a-4af3-82d5-800c1f80cba7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '722aa42e-f99c-42a6-aebc-f5ea99447e2f', 'attached_at': '2024-10-20T00:40:28.000000', 'detached_at': '', 'volume_id': '0796a138-bf1a-4af3-82d5-800c1f80cba7', 'serial': '0796a138-bf1a-4af3-82d5-800c1f80cba7'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1152.099411] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfc0250-c94c-4798-9e25-444c06604d64 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.128545] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fadff81-1e5a-44ca-a200-f0ecc2f3b614 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.137149] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f714f92-89da-458c-bf25-43d2431fc3b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.158346] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fab1d0-dcbb-4630-b7ad-4d69dac82f09 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.176009] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] The volume has not been displaced from its original location: [datastore1] volume-0796a138-bf1a-4af3-82d5-800c1f80cba7/volume-0796a138-bf1a-4af3-82d5-800c1f80cba7.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1152.181386] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1152.184537] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10e82daf-a6a3-4e2f-b744-94f75fbd0ce0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.203409] env[62521]: DEBUG oslo_vmware.api [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319276, 'name': PowerOnVM_Task, 'duration_secs': 0.49765} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.204782] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1152.205014] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-44070b4e-733f-4ee2-a886-f4718e521c76 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance '9eeeef50-4538-4b45-b14b-46671356adbf' progress to 100 {{(pid=62521) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1152.208786] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1152.208786] env[62521]: value = "task-1319277" [ 1152.208786] env[62521]: _type = "Task" [ 1152.208786] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.217320] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319277, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.301877] env[62521]: INFO nova.compute.manager [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Swapping old allocation on dict_keys(['3d21544b-bfc8-42d0-86ca-d323b5e2628f']) held by migration 140d6d52-68b2-4815-95a8-db67cb6f25b6 for instance [ 1152.315954] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1152.316126] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1152.323275] env[62521]: DEBUG nova.scheduler.client.report [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Overwriting current allocation {'allocations': {'3d21544b-bfc8-42d0-86ca-d323b5e2628f': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 157}}, 'project_id': '31c788848b0d47478564e53066e7c51a', 'user_id': '8a0e35b0e4574ed9afc4549d30daa210', 'consumer_generation': 1} on consumer e159807b-c7b0-4d2e-a7be-426ecaf01785 {{(pid=62521) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1152.403687] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.403887] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.404083] env[62521]: DEBUG nova.network.neutron [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1152.462857] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319274, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515583} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.463213] env[62521]: INFO nova.virt.vmwareapi.ds_util [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore2] 1bea570f-f2b2-4e9a-8dbb-418025402d3d/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk. [ 1152.463941] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504f4741-7a7e-4918-945b-3b7c7570642c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.492349] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 1bea570f-f2b2-4e9a-8dbb-418025402d3d/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1152.492674] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2bd6bc8-95bc-48e1-9093-a6b3c2462a03 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.512720] env[62521]: DEBUG nova.scheduler.client.report [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1152.515848] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1152.515848] env[62521]: value = "task-1319278" [ 1152.515848] env[62521]: _type = "Task" [ 1152.515848] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.525736] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319278, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.723997] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319277, 'name': ReconfigVM_Task, 'duration_secs': 0.213936} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.724331] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1152.729801] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93c05f18-b4fd-4d1a-a129-0b7a79f0552a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.749730] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1152.749730] env[62521]: value = "task-1319279" [ 1152.749730] env[62521]: _type = "Task" [ 1152.749730] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.759661] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319279, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.021126] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.021636] env[62521]: DEBUG nova.compute.manager [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1153.030695] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319278, 'name': ReconfigVM_Task, 'duration_secs': 0.351995} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.030994] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 1bea570f-f2b2-4e9a-8dbb-418025402d3d/58bd9a24-a0a4-4846-b299-475fa0f1d05d-rescue.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1153.031845] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab20d547-7c53-4f0a-8e64-803ed5139ce5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.062683] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b55904a5-4178-4cb1-8607-cb6b518aadf9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.078914] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1153.078914] env[62521]: value = "task-1319280" [ 1153.078914] env[62521]: _type = "Task" [ 1153.078914] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.087241] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319280, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.181151] env[62521]: DEBUG nova.network.neutron [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance_info_cache with network_info: [{"id": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "address": "fa:16:3e:20:00:55", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff3e4ad4-29", "ovs_interfaceid": "ff3e4ad4-29de-4ab1-bf26-bd8554ee6260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.260095] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319279, 'name': ReconfigVM_Task, 'duration_secs': 0.183774} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.260408] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282198', 'volume_id': '0796a138-bf1a-4af3-82d5-800c1f80cba7', 'name': 'volume-0796a138-bf1a-4af3-82d5-800c1f80cba7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '722aa42e-f99c-42a6-aebc-f5ea99447e2f', 'attached_at': '2024-10-20T00:40:28.000000', 'detached_at': '', 'volume_id': '0796a138-bf1a-4af3-82d5-800c1f80cba7', 'serial': '0796a138-bf1a-4af3-82d5-800c1f80cba7'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1153.260701] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1153.261635] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebefd4f-c126-4e02-8eec-ea5ac68f6a83 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.269362] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1153.269600] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c769f09b-42d5-4b23-94f7-8db4543acdff {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.311526] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.334801] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1153.335096] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1153.335396] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleting the datastore file [datastore1] 722aa42e-f99c-42a6-aebc-f5ea99447e2f {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.335719] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8d77060-2576-4d1b-8867-6409f0df6e8b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.342922] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1153.342922] env[62521]: value = "task-1319282" [ 1153.342922] env[62521]: _type = "Task" [ 1153.342922] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.351091] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.527492] env[62521]: DEBUG nova.compute.utils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1153.529504] env[62521]: DEBUG nova.compute.manager [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1153.529721] env[62521]: DEBUG nova.network.neutron [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1153.568014] env[62521]: DEBUG nova.policy [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ce6cad0be5f455eab09b1d348268329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '961924204a584b61a4a85a761821910d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1153.589304] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319280, 'name': ReconfigVM_Task, 'duration_secs': 0.183168} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.589586] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1153.589898] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3fc13d19-6030-44b1-9dcd-739e30fdd57f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.598297] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1153.598297] env[62521]: value = "task-1319283" [ 1153.598297] env[62521]: _type = "Task" [ 1153.598297] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.606341] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319283, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.684068] env[62521]: DEBUG oslo_concurrency.lockutils [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-e159807b-c7b0-4d2e-a7be-426ecaf01785" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.684548] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1153.684833] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f305ee3-24ee-45ed-95b5-7daf9b54c757 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.692829] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1153.692829] env[62521]: value = "task-1319284" [ 1153.692829] env[62521]: _type = "Task" [ 1153.692829] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.701824] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319284, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.855438] env[62521]: DEBUG oslo_vmware.api [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319282, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300808} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.855867] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1153.856147] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1153.856387] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1153.856612] env[62521]: INFO nova.compute.manager [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Took 2.28 seconds to destroy the instance on the hypervisor. [ 1153.856902] env[62521]: DEBUG oslo.service.loopingcall [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1153.857160] env[62521]: DEBUG nova.compute.manager [-] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1153.857430] env[62521]: DEBUG nova.network.neutron [-] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1154.032152] env[62521]: DEBUG nova.network.neutron [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Successfully created port: fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1154.035135] env[62521]: DEBUG nova.compute.manager [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1154.111063] env[62521]: DEBUG oslo_vmware.api [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319283, 'name': PowerOnVM_Task, 'duration_secs': 0.421979} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.111500] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1154.114316] env[62521]: DEBUG nova.compute.manager [None req-772266be-ac63-4e51-b3ff-bc50d5be79bf tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1154.115339] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23a79bf-cbc9-4d2c-9d3e-3cc945f98caf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.204542] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319284, 'name': PowerOffVM_Task, 'duration_secs': 0.26003} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.204834] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1154.205595] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1154.205802] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1154.205970] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1154.206181] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1154.206335] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1154.206490] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1154.206809] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1154.206898] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1154.207057] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1154.207263] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1154.211568] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1154.218854] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-441fadbd-c650-4d21-a51e-e76c3bb64a78 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.236984] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1154.236984] env[62521]: value = "task-1319285" [ 1154.236984] env[62521]: _type = "Task" [ 1154.236984] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.248638] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319285, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.318988] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1154.487407] env[62521]: DEBUG nova.compute.manager [req-ea12e697-0177-4692-a72b-4a65ab0e6e82 req-de2a127b-4aba-4530-982a-96195c6d533e service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Received event network-vif-deleted-c6383189-f73f-4f87-8424-e8956b34f94c {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1154.487407] env[62521]: INFO nova.compute.manager [req-ea12e697-0177-4692-a72b-4a65ab0e6e82 req-de2a127b-4aba-4530-982a-96195c6d533e service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Neutron deleted interface c6383189-f73f-4f87-8424-e8956b34f94c; detaching it from the instance and deleting it from the info cache [ 1154.487407] env[62521]: DEBUG nova.network.neutron [req-ea12e697-0177-4692-a72b-4a65ab0e6e82 req-de2a127b-4aba-4530-982a-96195c6d533e service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.749131] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319285, 'name': ReconfigVM_Task, 'duration_secs': 0.193793} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.750112] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198e21c9-fa9a-4cfe-ae1f-6a429e670c7b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.780794] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1154.781190] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1154.781449] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1154.781735] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1154.781978] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1154.782239] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1154.782554] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1154.782810] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1154.783095] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1154.783365] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1154.783653] env[62521]: DEBUG nova.virt.hardware [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1154.784901] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eea743a9-c661-45ca-a4b0-c9002d1fd65c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.793673] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1154.793673] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52569cfe-5313-333f-07e4-08a22fa6c20c" [ 1154.793673] env[62521]: _type = "Task" [ 1154.793673] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.804130] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52569cfe-5313-333f-07e4-08a22fa6c20c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.821147] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "9eeeef50-4538-4b45-b14b-46671356adbf" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.821270] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "9eeeef50-4538-4b45-b14b-46671356adbf" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.821457] env[62521]: DEBUG nova.compute.manager [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Going to confirm migration 9 {{(pid=62521) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1154.953941] env[62521]: DEBUG nova.network.neutron [-] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.992037] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0d36f275-25a8-4b14-bd21-1ecddf89032b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.003128] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09fbfa71-1ffd-4ec8-a170-7d9d6e3df743 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.037494] env[62521]: DEBUG nova.compute.manager [req-ea12e697-0177-4692-a72b-4a65ab0e6e82 req-de2a127b-4aba-4530-982a-96195c6d533e service nova] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Detach interface failed, port_id=c6383189-f73f-4f87-8424-e8956b34f94c, reason: Instance 722aa42e-f99c-42a6-aebc-f5ea99447e2f could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1155.048244] env[62521]: DEBUG nova.compute.manager [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1155.111749] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1155.112028] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1155.112311] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1155.112522] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1155.112804] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1155.112986] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1155.113228] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1155.113391] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1155.113565] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1155.113740] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1155.113922] env[62521]: DEBUG nova.virt.hardware [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1155.114810] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575df04f-e10e-40c7-ba7e-cf2299edea98 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.123949] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632d06c4-ecd9-44ec-8f98-867333f4145b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.306678] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52569cfe-5313-333f-07e4-08a22fa6c20c, 'name': SearchDatastore_Task, 'duration_secs': 0.026897} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.312445] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1155.312784] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2551cd66-3716-4652-9f7d-2b05e314e3ad {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.325926] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1155.326094] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1155.337772] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1155.337772] env[62521]: value = "task-1319286" [ 1155.337772] env[62521]: _type = "Task" [ 1155.337772] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.350982] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319286, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.457404] env[62521]: INFO nova.compute.manager [-] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Took 1.60 seconds to deallocate network for instance. [ 1155.503716] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.503716] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquired lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.503716] env[62521]: DEBUG nova.network.neutron [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1155.503910] env[62521]: DEBUG nova.objects.instance [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lazy-loading 'info_cache' on Instance uuid 9eeeef50-4538-4b45-b14b-46671356adbf {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1155.851659] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319286, 'name': ReconfigVM_Task, 'duration_secs': 0.42512} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.851659] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1155.851659] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7088be7a-2740-4957-a56f-6582fc3a3a50 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.877375] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] e159807b-c7b0-4d2e-a7be-426ecaf01785/e159807b-c7b0-4d2e-a7be-426ecaf01785.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1155.878048] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62b85b31-8ea1-4112-b900-b24844fdb153 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.898822] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1155.898822] env[62521]: value = "task-1319287" [ 1155.898822] env[62521]: _type = "Task" [ 1155.898822] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.899959] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.900243] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquired lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.900811] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Forcefully refreshing network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1155.913479] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319287, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.943206] env[62521]: DEBUG nova.network.neutron [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Successfully updated port: fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1155.970861] env[62521]: DEBUG nova.compute.manager [req-7808f46d-03b6-4dfe-9466-17ac63f02cdd req-47a5f2dc-88bf-41a2-8c93-40ce76ffaeac service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received event network-vif-plugged-fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1155.970988] env[62521]: DEBUG oslo_concurrency.lockutils [req-7808f46d-03b6-4dfe-9466-17ac63f02cdd req-47a5f2dc-88bf-41a2-8c93-40ce76ffaeac service nova] Acquiring lock "144636f8-8650-4448-8fe5-aeec0ba6b69e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.971256] env[62521]: DEBUG oslo_concurrency.lockutils [req-7808f46d-03b6-4dfe-9466-17ac63f02cdd req-47a5f2dc-88bf-41a2-8c93-40ce76ffaeac service nova] Lock "144636f8-8650-4448-8fe5-aeec0ba6b69e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.971425] env[62521]: DEBUG oslo_concurrency.lockutils [req-7808f46d-03b6-4dfe-9466-17ac63f02cdd req-47a5f2dc-88bf-41a2-8c93-40ce76ffaeac service nova] Lock "144636f8-8650-4448-8fe5-aeec0ba6b69e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.971707] env[62521]: DEBUG nova.compute.manager [req-7808f46d-03b6-4dfe-9466-17ac63f02cdd req-47a5f2dc-88bf-41a2-8c93-40ce76ffaeac service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] No waiting events found dispatching network-vif-plugged-fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1155.971820] env[62521]: WARNING nova.compute.manager [req-7808f46d-03b6-4dfe-9466-17ac63f02cdd req-47a5f2dc-88bf-41a2-8c93-40ce76ffaeac service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received unexpected event network-vif-plugged-fe1d7288-4afc-44de-a9f7-4e98766784d4 for instance with vm_state building and task_state spawning. [ 1156.019375] env[62521]: INFO nova.compute.manager [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 722aa42e-f99c-42a6-aebc-f5ea99447e2f] Took 0.56 seconds to detach 1 volumes for instance. [ 1156.417218] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319287, 'name': ReconfigVM_Task, 'duration_secs': 0.35031} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.417548] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Reconfigured VM instance instance-00000069 to attach disk [datastore1] e159807b-c7b0-4d2e-a7be-426ecaf01785/e159807b-c7b0-4d2e-a7be-426ecaf01785.vmdk or device None with type thin {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1156.418476] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b82b6f4-10c3-4b06-a7a1-af4ea427ef1b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.438677] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b18e14-1bcc-419f-b514-5c98f0b4378f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.457888] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.458626] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.458626] env[62521]: DEBUG nova.network.neutron [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1156.462556] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fe0054-1dc0-4737-ac38-01dd995c4738 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.482851] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd84c31a-f41e-43da-9eae-3dddec0bd6f7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.491433] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1156.491718] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00f61b56-ae8e-42b3-a1c5-79dfa9948a20 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.498826] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1156.498826] env[62521]: value = "task-1319288" [ 1156.498826] env[62521]: _type = "Task" [ 1156.498826] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.508498] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319288, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.528585] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.528918] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.529129] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.538302] env[62521]: INFO nova.compute.manager [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Unrescuing [ 1156.538302] env[62521]: DEBUG oslo_concurrency.lockutils [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.538302] env[62521]: DEBUG oslo_concurrency.lockutils [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquired lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.538564] env[62521]: DEBUG nova.network.neutron [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1156.557139] env[62521]: INFO nova.scheduler.client.report [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleted allocations for instance 722aa42e-f99c-42a6-aebc-f5ea99447e2f [ 1156.775582] env[62521]: DEBUG nova.network.neutron [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance_info_cache with network_info: [{"id": "74195062-c22d-4315-8ff8-47c98a866c60", "address": "fa:16:3e:29:21:07", "network": {"id": "b0a61364-dfa9-4e50-ae33-a354c077b517", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-90311723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0dfb5f5bd4646ec95b7c6a2d0434175", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74195062-c2", "ovs_interfaceid": "74195062-c22d-4315-8ff8-47c98a866c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.995802] env[62521]: DEBUG nova.network.neutron [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1157.011872] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319288, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.066095] env[62521]: DEBUG oslo_concurrency.lockutils [None req-4821eee1-9706-4836-89d7-f975c26aa029 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "722aa42e-f99c-42a6-aebc-f5ea99447e2f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.492s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.139165] env[62521]: DEBUG nova.network.neutron [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Updating instance_info_cache with network_info: [{"id": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "address": "fa:16:3e:44:95:f8", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee2404e8-7a", "ovs_interfaceid": "ee2404e8-7a05-4516-826e-2f5e4adcae45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.278905] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Releasing lock "refresh_cache-9eeeef50-4538-4b45-b14b-46671356adbf" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.279219] env[62521]: DEBUG nova.objects.instance [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lazy-loading 'migration_context' on Instance uuid 9eeeef50-4538-4b45-b14b-46671356adbf {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.297679] env[62521]: DEBUG nova.network.neutron [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updating instance_info_cache with network_info: [{"id": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "address": "fa:16:3e:08:82:2a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1d7288-4a", "ovs_interfaceid": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.404786] env[62521]: DEBUG nova.network.neutron [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updating instance_info_cache with network_info: [{"id": "02aabc96-04f0-4b09-bcf6-44443d6542af", "address": "fa:16:3e:4a:64:63", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aabc96-04", "ovs_interfaceid": "02aabc96-04f0-4b09-bcf6-44443d6542af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.510099] env[62521]: DEBUG oslo_vmware.api [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319288, 'name': PowerOnVM_Task, 'duration_secs': 0.799877} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.510411] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1157.641872] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Releasing lock "refresh_cache-703ff423-98b1-4942-b6fd-8b95fe57bd0a" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.642116] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Updated the network info_cache for instance {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1157.642795] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1157.642795] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1157.642795] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1157.642795] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1157.643044] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1157.782542] env[62521]: DEBUG nova.objects.base [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Object Instance<9eeeef50-4538-4b45-b14b-46671356adbf> lazy-loaded attributes: info_cache,migration_context {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1157.783502] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df56b00-a6b5-41ea-b43c-7ca1d884ae28 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.802975] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.803286] env[62521]: DEBUG nova.compute.manager [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Instance network_info: |[{"id": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "address": "fa:16:3e:08:82:2a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1d7288-4a", "ovs_interfaceid": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1157.804263] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:82:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92cdccfd-4b10-4024-b724-5f22792dd4de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe1d7288-4afc-44de-a9f7-4e98766784d4', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1157.811425] env[62521]: DEBUG oslo.service.loopingcall [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1157.811610] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a46f9c6-20f8-4844-8956-24800084daba {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.814139] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1157.814352] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-90504e72-ec1d-46d1-aafe-fa072bfa2785 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.833582] env[62521]: DEBUG oslo_vmware.api [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1157.833582] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525cb5e9-c7f7-5d5e-f216-b346b59971d9" [ 1157.833582] env[62521]: _type = "Task" [ 1157.833582] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.838308] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1157.838308] env[62521]: value = "task-1319289" [ 1157.838308] env[62521]: _type = "Task" [ 1157.838308] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.845379] env[62521]: DEBUG oslo_vmware.api [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]525cb5e9-c7f7-5d5e-f216-b346b59971d9, 'name': SearchDatastore_Task} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.846102] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.846417] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.851262] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319289, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.908506] env[62521]: DEBUG oslo_concurrency.lockutils [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Releasing lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.909210] env[62521]: DEBUG nova.objects.instance [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lazy-loading 'flavor' on Instance uuid 1bea570f-f2b2-4e9a-8dbb-418025402d3d {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.996894] env[62521]: DEBUG nova.compute.manager [req-7aa4f49d-3ce6-45d1-8f4d-7d8691c18a5a req-ce4cdf73-1b0a-4e0b-acf0-c07272017b46 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received event network-changed-fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1157.997166] env[62521]: DEBUG nova.compute.manager [req-7aa4f49d-3ce6-45d1-8f4d-7d8691c18a5a req-ce4cdf73-1b0a-4e0b-acf0-c07272017b46 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing instance network info cache due to event network-changed-fe1d7288-4afc-44de-a9f7-4e98766784d4. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1157.997380] env[62521]: DEBUG oslo_concurrency.lockutils [req-7aa4f49d-3ce6-45d1-8f4d-7d8691c18a5a req-ce4cdf73-1b0a-4e0b-acf0-c07272017b46 service nova] Acquiring lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.997554] env[62521]: DEBUG oslo_concurrency.lockutils [req-7aa4f49d-3ce6-45d1-8f4d-7d8691c18a5a req-ce4cdf73-1b0a-4e0b-acf0-c07272017b46 service nova] Acquired lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.997724] env[62521]: DEBUG nova.network.neutron [req-7aa4f49d-3ce6-45d1-8f4d-7d8691c18a5a req-ce4cdf73-1b0a-4e0b-acf0-c07272017b46 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing network info cache for port fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1158.145492] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.215687] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.215687] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.349504] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319289, 'name': CreateVM_Task, 'duration_secs': 0.349294} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.349925] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1158.350726] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.351098] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.351524] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1158.354104] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb2dad6e-b334-41a2-84b0-cda90b265431 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.360440] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1158.360440] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5286d626-8aa1-669d-a3cd-c0eb1d0a0d33" [ 1158.360440] env[62521]: _type = "Task" [ 1158.360440] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.368419] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5286d626-8aa1-669d-a3cd-c0eb1d0a0d33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.416792] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f2b581-8dfd-4de4-952e-44c524bc3ed2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.440827] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1158.443354] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d58cf0f-f638-470a-b30f-bda591fd7cb7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.450535] env[62521]: DEBUG oslo_vmware.api [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1158.450535] env[62521]: value = "task-1319290" [ 1158.450535] env[62521]: _type = "Task" [ 1158.450535] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.463296] env[62521]: DEBUG oslo_vmware.api [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319290, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.476133] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d933a4-bfcc-42c7-94dd-abf044f6d79c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.484663] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f794afb4-30fa-4edd-8fa2-db63dff517bb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.517261] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d4cb98-0e09-4680-9048-80b100d517d4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.521115] env[62521]: INFO nova.compute.manager [None req-ab0ecbe1-abbb-48c4-be34-86cfe3e2fad9 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance to original state: 'active' [ 1158.529954] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdae12c-1bbb-4ac4-ade2-42245fbfcf53 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.544455] env[62521]: DEBUG nova.compute.provider_tree [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.717382] env[62521]: DEBUG nova.compute.manager [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1158.741224] env[62521]: DEBUG nova.network.neutron [req-7aa4f49d-3ce6-45d1-8f4d-7d8691c18a5a req-ce4cdf73-1b0a-4e0b-acf0-c07272017b46 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updated VIF entry in instance network info cache for port fe1d7288-4afc-44de-a9f7-4e98766784d4. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1158.741683] env[62521]: DEBUG nova.network.neutron [req-7aa4f49d-3ce6-45d1-8f4d-7d8691c18a5a req-ce4cdf73-1b0a-4e0b-acf0-c07272017b46 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updating instance_info_cache with network_info: [{"id": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "address": "fa:16:3e:08:82:2a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1d7288-4a", "ovs_interfaceid": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.871403] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5286d626-8aa1-669d-a3cd-c0eb1d0a0d33, 'name': SearchDatastore_Task, 'duration_secs': 0.010009} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.871719] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.871957] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1158.872217] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.872368] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.872551] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1158.872815] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f66bbafd-70cf-44a0-9c5c-97778a8005ed {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.881779] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1158.881968] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1158.882760] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-560778a1-19d4-479f-9587-3867ead03cae {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.888060] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1158.888060] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524f3bc9-6305-e90f-7f7f-83ea4cd70b44" [ 1158.888060] env[62521]: _type = "Task" [ 1158.888060] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.896596] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524f3bc9-6305-e90f-7f7f-83ea4cd70b44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.963020] env[62521]: DEBUG oslo_vmware.api [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319290, 'name': PowerOffVM_Task, 'duration_secs': 0.261251} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.963367] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1158.968604] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Reconfiguring VM instance instance-0000006a to detach disk 2002 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1158.968958] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fdc8305-07cb-4910-a5d6-71ba64243027 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.996932] env[62521]: DEBUG oslo_vmware.api [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1158.996932] env[62521]: value = "task-1319291" [ 1158.996932] env[62521]: _type = "Task" [ 1158.996932] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.005318] env[62521]: DEBUG oslo_vmware.api [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319291, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.047758] env[62521]: DEBUG nova.scheduler.client.report [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1159.236783] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.245798] env[62521]: DEBUG oslo_concurrency.lockutils [req-7aa4f49d-3ce6-45d1-8f4d-7d8691c18a5a req-ce4cdf73-1b0a-4e0b-acf0-c07272017b46 service nova] Releasing lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.398704] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524f3bc9-6305-e90f-7f7f-83ea4cd70b44, 'name': SearchDatastore_Task, 'duration_secs': 0.009148} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.399474] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-596b4ff4-40ad-4891-b21e-93cab3d960d9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.405116] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1159.405116] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5265f6f4-c95a-4803-376b-c0376cc0d7d9" [ 1159.405116] env[62521]: _type = "Task" [ 1159.405116] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.413096] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5265f6f4-c95a-4803-376b-c0376cc0d7d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.507966] env[62521]: DEBUG oslo_vmware.api [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319291, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.849525] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "e159807b-c7b0-4d2e-a7be-426ecaf01785" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.849722] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.850086] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "e159807b-c7b0-4d2e-a7be-426ecaf01785-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.850345] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.850570] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.852767] env[62521]: INFO nova.compute.manager [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Terminating instance [ 1159.854555] env[62521]: DEBUG nova.compute.manager [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1159.854789] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1159.855702] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97717ca-b9e0-4b2e-bfbc-845941729705 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.864359] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1159.864602] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a51f6156-e23f-4807-b262-80ed312ce1fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.872090] env[62521]: DEBUG oslo_vmware.api [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1159.872090] env[62521]: value = "task-1319292" [ 1159.872090] env[62521]: _type = "Task" [ 1159.872090] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.881767] env[62521]: DEBUG oslo_vmware.api [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319292, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.916120] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5265f6f4-c95a-4803-376b-c0376cc0d7d9, 'name': SearchDatastore_Task, 'duration_secs': 0.027186} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.916416] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.916685] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 144636f8-8650-4448-8fe5-aeec0ba6b69e/144636f8-8650-4448-8fe5-aeec0ba6b69e.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1159.916997] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee3c71bf-611c-42e8-8791-dce8276d2d40 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.924819] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1159.924819] env[62521]: value = "task-1319293" [ 1159.924819] env[62521]: _type = "Task" [ 1159.924819] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.933547] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319293, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.008498] env[62521]: DEBUG oslo_vmware.api [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319291, 'name': ReconfigVM_Task, 'duration_secs': 0.713485} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.008929] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Reconfigured VM instance instance-0000006a to detach disk 2002 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1160.008981] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1160.009312] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b19e41cd-a9e5-4268-94d1-eee7dda69769 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.018943] env[62521]: DEBUG oslo_vmware.api [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1160.018943] env[62521]: value = "task-1319294" [ 1160.018943] env[62521]: _type = "Task" [ 1160.018943] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.029647] env[62521]: DEBUG oslo_vmware.api [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319294, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.057755] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.211s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.061284] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.916s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.061484] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.061667] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1160.062094] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.825s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.064316] env[62521]: INFO nova.compute.claims [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1160.067580] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834acccc-6532-4a72-a506-9061594d0301 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.077816] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4652a7-241a-4ecd-a55c-a099a14fa6d6 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.093211] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb683c8-ecdc-4e77-9268-b5687ec402bf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.102226] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b9d759-d112-49ef-8a30-067542ec8889 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.133968] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179988MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1160.134170] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.382502] env[62521]: DEBUG oslo_vmware.api [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319292, 'name': PowerOffVM_Task, 'duration_secs': 0.169734} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.382802] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1160.382980] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1160.383278] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2893e092-db35-4340-a3b3-db020190c68d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.436470] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319293, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.529227] env[62521]: DEBUG oslo_vmware.api [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319294, 'name': PowerOnVM_Task, 'duration_secs': 0.423261} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.529519] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1160.529758] env[62521]: DEBUG nova.compute.manager [None req-434b3d96-2e29-43ec-ad02-f1a52e1c5682 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1160.530583] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8d68c9-0efe-456d-8340-e5cb86548729 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.628999] env[62521]: INFO nova.scheduler.client.report [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted allocation for migration 90675f29-1342-4dfb-b872-f26c26f1cfe4 [ 1160.938490] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319293, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527876} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.938670] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 144636f8-8650-4448-8fe5-aeec0ba6b69e/144636f8-8650-4448-8fe5-aeec0ba6b69e.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1160.938866] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1160.939147] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-688533d6-2ae9-4b57-8abd-57312ddb5f7b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.948642] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1160.948642] env[62521]: value = "task-1319296" [ 1160.948642] env[62521]: _type = "Task" [ 1160.948642] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.957126] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319296, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.122928] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1161.123247] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1161.123371] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleting the datastore file [datastore1] e159807b-c7b0-4d2e-a7be-426ecaf01785 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1161.123640] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a44f5707-ab34-452c-bce9-241e9757dfc2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.130285] env[62521]: DEBUG oslo_vmware.api [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1161.130285] env[62521]: value = "task-1319297" [ 1161.130285] env[62521]: _type = "Task" [ 1161.130285] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.134472] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "9eeeef50-4538-4b45-b14b-46671356adbf" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.313s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.143974] env[62521]: DEBUG oslo_vmware.api [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319297, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.176426] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e5a568-ce7a-456b-8e21-3adb2e90dca4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.184634] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3fec0d-487d-44bb-b258-1495bf7ae72b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.215213] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb0dbd9-e467-4449-bdd6-d8fd4519c01f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.223358] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b121acf-d4f5-4fc1-a4b0-8ad9f74822f1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.238612] env[62521]: DEBUG nova.compute.provider_tree [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1161.458555] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319296, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.253762} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.458841] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1161.459615] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0759d306-7295-49b6-9ad9-5eac0710d3ef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.481530] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 144636f8-8650-4448-8fe5-aeec0ba6b69e/144636f8-8650-4448-8fe5-aeec0ba6b69e.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1161.481817] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e86610d7-8c17-4f34-9117-8072e3ddafb5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.501393] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1161.501393] env[62521]: value = "task-1319298" [ 1161.501393] env[62521]: _type = "Task" [ 1161.501393] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.509222] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319298, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.623154] env[62521]: DEBUG nova.compute.manager [req-626d5378-b362-49e4-be5b-283892efbbfa req-d0d35815-e4f6-4af1-ab78-bd2f254a20ea service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Received event network-changed-02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1161.623599] env[62521]: DEBUG nova.compute.manager [req-626d5378-b362-49e4-be5b-283892efbbfa req-d0d35815-e4f6-4af1-ab78-bd2f254a20ea service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Refreshing instance network info cache due to event network-changed-02aabc96-04f0-4b09-bcf6-44443d6542af. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1161.623844] env[62521]: DEBUG oslo_concurrency.lockutils [req-626d5378-b362-49e4-be5b-283892efbbfa req-d0d35815-e4f6-4af1-ab78-bd2f254a20ea service nova] Acquiring lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.624075] env[62521]: DEBUG oslo_concurrency.lockutils [req-626d5378-b362-49e4-be5b-283892efbbfa req-d0d35815-e4f6-4af1-ab78-bd2f254a20ea service nova] Acquired lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.624334] env[62521]: DEBUG nova.network.neutron [req-626d5378-b362-49e4-be5b-283892efbbfa req-d0d35815-e4f6-4af1-ab78-bd2f254a20ea service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Refreshing network info cache for port 02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1161.641207] env[62521]: DEBUG oslo_vmware.api [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319297, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160486} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.641486] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1161.641672] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1161.641859] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1161.642045] env[62521]: INFO nova.compute.manager [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Took 1.79 seconds to destroy the instance on the hypervisor. [ 1161.642294] env[62521]: DEBUG oslo.service.loopingcall [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1161.642485] env[62521]: DEBUG nova.compute.manager [-] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1161.642581] env[62521]: DEBUG nova.network.neutron [-] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1161.742041] env[62521]: DEBUG nova.scheduler.client.report [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1162.012394] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319298, 'name': ReconfigVM_Task, 'duration_secs': 0.297948} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.012572] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 144636f8-8650-4448-8fe5-aeec0ba6b69e/144636f8-8650-4448-8fe5-aeec0ba6b69e.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1162.013382] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92281ce1-4f40-46c6-9e3e-ef5d5015ad57 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.023484] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1162.023484] env[62521]: value = "task-1319299" [ 1162.023484] env[62521]: _type = "Task" [ 1162.023484] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.033441] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319299, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.247055] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.185s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.247664] env[62521]: DEBUG nova.compute.manager [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1162.250633] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.116s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.342035] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "9eeeef50-4538-4b45-b14b-46671356adbf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.342182] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "9eeeef50-4538-4b45-b14b-46671356adbf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.342398] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "9eeeef50-4538-4b45-b14b-46671356adbf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.342590] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "9eeeef50-4538-4b45-b14b-46671356adbf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.342760] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "9eeeef50-4538-4b45-b14b-46671356adbf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.344985] env[62521]: INFO nova.compute.manager [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Terminating instance [ 1162.346906] env[62521]: DEBUG nova.compute.manager [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1162.346994] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1162.349827] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d87b71-2f0f-4477-b2c9-fa55d1c0f614 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.357206] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1162.357961] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4430213a-2d66-446d-8c98-441df69713e2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.365037] env[62521]: DEBUG oslo_vmware.api [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1162.365037] env[62521]: value = "task-1319300" [ 1162.365037] env[62521]: _type = "Task" [ 1162.365037] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.375317] env[62521]: DEBUG oslo_vmware.api [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319300, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.407069] env[62521]: DEBUG nova.network.neutron [req-626d5378-b362-49e4-be5b-283892efbbfa req-d0d35815-e4f6-4af1-ab78-bd2f254a20ea service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updated VIF entry in instance network info cache for port 02aabc96-04f0-4b09-bcf6-44443d6542af. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1162.407474] env[62521]: DEBUG nova.network.neutron [req-626d5378-b362-49e4-be5b-283892efbbfa req-d0d35815-e4f6-4af1-ab78-bd2f254a20ea service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updating instance_info_cache with network_info: [{"id": "02aabc96-04f0-4b09-bcf6-44443d6542af", "address": "fa:16:3e:4a:64:63", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aabc96-04", "ovs_interfaceid": "02aabc96-04f0-4b09-bcf6-44443d6542af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.531875] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319299, 'name': Rename_Task, 'duration_secs': 0.153534} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.532237] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1162.532525] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d51272e-8963-4673-b7fb-32e06861a191 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.539915] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1162.539915] env[62521]: value = "task-1319301" [ 1162.539915] env[62521]: _type = "Task" [ 1162.539915] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.550279] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319301, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.684701] env[62521]: DEBUG nova.network.neutron [-] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.754869] env[62521]: DEBUG nova.compute.utils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1162.762253] env[62521]: DEBUG nova.compute.manager [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1162.762509] env[62521]: DEBUG nova.network.neutron [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1162.804020] env[62521]: DEBUG nova.policy [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfab7121825d4e049370b526f90eb499', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11829634203c4967bbcff90eb904097c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1162.877382] env[62521]: DEBUG oslo_vmware.api [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319300, 'name': PowerOffVM_Task, 'duration_secs': 0.212138} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.877662] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1162.877881] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1162.878117] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-243a70ac-6adf-4366-9703-309539a81ac9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.910267] env[62521]: DEBUG oslo_concurrency.lockutils [req-626d5378-b362-49e4-be5b-283892efbbfa req-d0d35815-e4f6-4af1-ab78-bd2f254a20ea service nova] Releasing lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.950282] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1162.950651] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1162.951065] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleting the datastore file [datastore1] 9eeeef50-4538-4b45-b14b-46671356adbf {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1162.951362] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e7f9978-41ba-4931-a208-ea1de181e63c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.960543] env[62521]: DEBUG oslo_vmware.api [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for the task: (returnval){ [ 1162.960543] env[62521]: value = "task-1319303" [ 1162.960543] env[62521]: _type = "Task" [ 1162.960543] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.973231] env[62521]: DEBUG oslo_vmware.api [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319303, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.051746] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319301, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.148056] env[62521]: DEBUG nova.network.neutron [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Successfully created port: c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1163.187798] env[62521]: INFO nova.compute.manager [-] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Took 1.55 seconds to deallocate network for instance. [ 1163.265793] env[62521]: DEBUG nova.compute.manager [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1163.297612] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 703ff423-98b1-4942-b6fd-8b95fe57bd0a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1163.297779] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 4021ce0d-69ee-4c57-bb18-fd77f339ed72 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1163.297984] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 1bea570f-f2b2-4e9a-8dbb-418025402d3d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1163.298115] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 9eeeef50-4538-4b45-b14b-46671356adbf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1163.298234] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 144636f8-8650-4448-8fe5-aeec0ba6b69e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1163.298353] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance e159807b-c7b0-4d2e-a7be-426ecaf01785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1163.298469] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1163.298661] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1163.298829] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1163.418725] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045178e8-3377-4b26-bd83-26e19951d778 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.427647] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5950c0fa-515d-4e85-97d2-b3bbbac9d199 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.459799] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a26065-f6e8-4b29-8759-100cb2e7c985 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.474832] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d15bba7-5871-45eb-93c9-dfbbac19bd46 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.479216] env[62521]: DEBUG oslo_vmware.api [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Task: {'id': task-1319303, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172095} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.479494] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1163.479684] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1163.479866] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1163.480125] env[62521]: INFO nova.compute.manager [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1163.480380] env[62521]: DEBUG oslo.service.loopingcall [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1163.480987] env[62521]: DEBUG nova.compute.manager [-] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1163.481183] env[62521]: DEBUG nova.network.neutron [-] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1163.491819] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1163.551301] env[62521]: DEBUG oslo_vmware.api [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319301, 'name': PowerOnVM_Task, 'duration_secs': 0.5142} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.551587] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1163.551801] env[62521]: INFO nova.compute.manager [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Took 8.50 seconds to spawn the instance on the hypervisor. [ 1163.551986] env[62521]: DEBUG nova.compute.manager [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1163.552831] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f43e77-55e9-46d3-97c3-ec2a41fd12be {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.664187] env[62521]: DEBUG nova.compute.manager [req-925b762e-ec0d-40b2-af1f-84a0e374b67a req-be3c00ea-cede-44e0-82df-baa985a1497b service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Received event network-changed-02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1163.664402] env[62521]: DEBUG nova.compute.manager [req-925b762e-ec0d-40b2-af1f-84a0e374b67a req-be3c00ea-cede-44e0-82df-baa985a1497b service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Refreshing instance network info cache due to event network-changed-02aabc96-04f0-4b09-bcf6-44443d6542af. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1163.664629] env[62521]: DEBUG oslo_concurrency.lockutils [req-925b762e-ec0d-40b2-af1f-84a0e374b67a req-be3c00ea-cede-44e0-82df-baa985a1497b service nova] Acquiring lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.664830] env[62521]: DEBUG oslo_concurrency.lockutils [req-925b762e-ec0d-40b2-af1f-84a0e374b67a req-be3c00ea-cede-44e0-82df-baa985a1497b service nova] Acquired lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.664994] env[62521]: DEBUG nova.network.neutron [req-925b762e-ec0d-40b2-af1f-84a0e374b67a req-be3c00ea-cede-44e0-82df-baa985a1497b service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Refreshing network info cache for port 02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1163.694710] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.995026] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1164.077116] env[62521]: INFO nova.compute.manager [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Took 16.09 seconds to build instance. [ 1164.277245] env[62521]: DEBUG nova.compute.manager [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1164.308247] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1164.308523] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1164.308684] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1164.308875] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1164.309040] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1164.309203] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1164.309432] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1164.309614] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1164.309876] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1164.309959] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1164.310230] env[62521]: DEBUG nova.virt.hardware [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1164.311191] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b23e93-2fdc-4701-a596-fcf75aa12374 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.319839] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d927599-3cae-4a4c-8aa2-7fbca86fff14 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.421095] env[62521]: DEBUG nova.network.neutron [req-925b762e-ec0d-40b2-af1f-84a0e374b67a req-be3c00ea-cede-44e0-82df-baa985a1497b service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updated VIF entry in instance network info cache for port 02aabc96-04f0-4b09-bcf6-44443d6542af. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1164.421547] env[62521]: DEBUG nova.network.neutron [req-925b762e-ec0d-40b2-af1f-84a0e374b67a req-be3c00ea-cede-44e0-82df-baa985a1497b service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updating instance_info_cache with network_info: [{"id": "02aabc96-04f0-4b09-bcf6-44443d6542af", "address": "fa:16:3e:4a:64:63", "network": {"id": "f47d0e48-4f8d-4ac1-a620-fde5738fd17f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-486154479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d7ab38464a4c02a9dc94e04b4793ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02aabc96-04", "ovs_interfaceid": "02aabc96-04f0-4b09-bcf6-44443d6542af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.500069] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1164.500318] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.250s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.500599] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.806s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.500834] env[62521]: DEBUG nova.objects.instance [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'resources' on Instance uuid e159807b-c7b0-4d2e-a7be-426ecaf01785 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.542857] env[62521]: DEBUG nova.network.neutron [-] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.579146] env[62521]: DEBUG oslo_concurrency.lockutils [None req-8a13682d-d728-4bd4-b1fa-9703050a0bd7 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.602s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.924983] env[62521]: DEBUG oslo_concurrency.lockutils [req-925b762e-ec0d-40b2-af1f-84a0e374b67a req-be3c00ea-cede-44e0-82df-baa985a1497b service nova] Releasing lock "refresh_cache-1bea570f-f2b2-4e9a-8dbb-418025402d3d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.925301] env[62521]: DEBUG nova.compute.manager [req-925b762e-ec0d-40b2-af1f-84a0e374b67a req-be3c00ea-cede-44e0-82df-baa985a1497b service nova] [instance: e159807b-c7b0-4d2e-a7be-426ecaf01785] Received event network-vif-deleted-ff3e4ad4-29de-4ab1-bf26-bd8554ee6260 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1164.946808] env[62521]: DEBUG nova.network.neutron [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Successfully updated port: c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1165.046251] env[62521]: INFO nova.compute.manager [-] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Took 1.57 seconds to deallocate network for instance. [ 1165.096292] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9733d5-7f17-4129-ba00-f159929be6fc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.104637] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468e85c0-5bb1-439a-bad1-81484cb08878 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.135899] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f83951-e6e5-48ff-9fc5-ce9cf0271aee {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.143690] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9824b867-c6d4-4474-bbea-18ca23ee3db3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.157364] env[62521]: DEBUG nova.compute.provider_tree [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1165.390149] env[62521]: DEBUG nova.compute.manager [req-f3f8fb36-c3e9-48c7-a05b-acdf0209998e req-39f1e192-b78d-4a38-83be-763e2055b839 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received event network-changed-fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1165.390418] env[62521]: DEBUG nova.compute.manager [req-f3f8fb36-c3e9-48c7-a05b-acdf0209998e req-39f1e192-b78d-4a38-83be-763e2055b839 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing instance network info cache due to event network-changed-fe1d7288-4afc-44de-a9f7-4e98766784d4. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1165.392920] env[62521]: DEBUG oslo_concurrency.lockutils [req-f3f8fb36-c3e9-48c7-a05b-acdf0209998e req-39f1e192-b78d-4a38-83be-763e2055b839 service nova] Acquiring lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.393124] env[62521]: DEBUG oslo_concurrency.lockutils [req-f3f8fb36-c3e9-48c7-a05b-acdf0209998e req-39f1e192-b78d-4a38-83be-763e2055b839 service nova] Acquired lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.393304] env[62521]: DEBUG nova.network.neutron [req-f3f8fb36-c3e9-48c7-a05b-acdf0209998e req-39f1e192-b78d-4a38-83be-763e2055b839 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing network info cache for port fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1165.449106] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.449291] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.449535] env[62521]: DEBUG nova.network.neutron [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1165.554497] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.660425] env[62521]: DEBUG nova.scheduler.client.report [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1165.694647] env[62521]: DEBUG nova.compute.manager [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] [instance: 9eeeef50-4538-4b45-b14b-46671356adbf] Received event network-vif-deleted-74195062-c22d-4315-8ff8-47c98a866c60 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1165.694881] env[62521]: DEBUG nova.compute.manager [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received event network-vif-plugged-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1165.695315] env[62521]: DEBUG oslo_concurrency.lockutils [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] Acquiring lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.695577] env[62521]: DEBUG oslo_concurrency.lockutils [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.695758] env[62521]: DEBUG oslo_concurrency.lockutils [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.695938] env[62521]: DEBUG nova.compute.manager [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] No waiting events found dispatching network-vif-plugged-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1165.696172] env[62521]: WARNING nova.compute.manager [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received unexpected event network-vif-plugged-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 for instance with vm_state building and task_state spawning. [ 1165.696390] env[62521]: DEBUG nova.compute.manager [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received event network-changed-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1165.696615] env[62521]: DEBUG nova.compute.manager [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Refreshing instance network info cache due to event network-changed-c6359cbd-f300-4344-8a1b-bed3f0dccdf2. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1165.696806] env[62521]: DEBUG oslo_concurrency.lockutils [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] Acquiring lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.995184] env[62521]: DEBUG nova.network.neutron [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1166.165440] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.665s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.168185] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.613s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.168185] env[62521]: DEBUG nova.objects.instance [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lazy-loading 'resources' on Instance uuid 9eeeef50-4538-4b45-b14b-46671356adbf {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.190861] env[62521]: INFO nova.scheduler.client.report [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleted allocations for instance e159807b-c7b0-4d2e-a7be-426ecaf01785 [ 1166.220023] env[62521]: DEBUG nova.network.neutron [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updating instance_info_cache with network_info: [{"id": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "address": "fa:16:3e:9f:8f:e8", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6359cbd-f3", "ovs_interfaceid": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.266026] env[62521]: DEBUG nova.network.neutron [req-f3f8fb36-c3e9-48c7-a05b-acdf0209998e req-39f1e192-b78d-4a38-83be-763e2055b839 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updated VIF entry in instance network info cache for port fe1d7288-4afc-44de-a9f7-4e98766784d4. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1166.266026] env[62521]: DEBUG nova.network.neutron [req-f3f8fb36-c3e9-48c7-a05b-acdf0209998e req-39f1e192-b78d-4a38-83be-763e2055b839 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updating instance_info_cache with network_info: [{"id": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "address": "fa:16:3e:08:82:2a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1d7288-4a", "ovs_interfaceid": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.362599] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.362898] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.699851] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91deaf62-6050-4a5d-9123-38cc4b522f5c tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "e159807b-c7b0-4d2e-a7be-426ecaf01785" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.850s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.722266] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.722575] env[62521]: DEBUG nova.compute.manager [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Instance network_info: |[{"id": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "address": "fa:16:3e:9f:8f:e8", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6359cbd-f3", "ovs_interfaceid": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1166.722885] env[62521]: DEBUG oslo_concurrency.lockutils [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] Acquired lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.723086] env[62521]: DEBUG nova.network.neutron [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Refreshing network info cache for port c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1166.724138] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:8f:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6359cbd-f300-4344-8a1b-bed3f0dccdf2', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1166.731598] env[62521]: DEBUG oslo.service.loopingcall [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1166.734784] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1166.735361] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-148036d8-1eda-4a28-ba1d-862e4eb013d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.758573] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1166.758573] env[62521]: value = "task-1319304" [ 1166.758573] env[62521]: _type = "Task" [ 1166.758573] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.769368] env[62521]: DEBUG oslo_concurrency.lockutils [req-f3f8fb36-c3e9-48c7-a05b-acdf0209998e req-39f1e192-b78d-4a38-83be-763e2055b839 service nova] Releasing lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.769696] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319304, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.796124] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01abf67-e728-42ba-83c6-9057856508ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.803610] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b917f76a-11d1-4dc4-b238-8c4043610369 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.837416] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915ee005-9b2c-4189-9241-a99cf7b7de67 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.853815] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1303426d-ca92-45eb-bf6f-6ebe2879634e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.867685] env[62521]: DEBUG nova.compute.manager [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1166.870712] env[62521]: DEBUG nova.compute.provider_tree [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.268876] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319304, 'name': CreateVM_Task, 'duration_secs': 0.390056} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.270998] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1167.271759] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.271930] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.272276] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1167.272800] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a23a108d-da1d-4844-8b0c-86aa909ce4c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.277585] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1167.277585] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5247bf1e-c2a7-f1c2-c668-47c8bc4935c3" [ 1167.277585] env[62521]: _type = "Task" [ 1167.277585] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.285987] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5247bf1e-c2a7-f1c2-c668-47c8bc4935c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.375071] env[62521]: DEBUG nova.scheduler.client.report [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1167.395943] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.484765] env[62521]: DEBUG nova.network.neutron [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updated VIF entry in instance network info cache for port c6359cbd-f300-4344-8a1b-bed3f0dccdf2. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1167.484765] env[62521]: DEBUG nova.network.neutron [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updating instance_info_cache with network_info: [{"id": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "address": "fa:16:3e:9f:8f:e8", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6359cbd-f3", "ovs_interfaceid": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.789618] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5247bf1e-c2a7-f1c2-c668-47c8bc4935c3, 'name': SearchDatastore_Task, 'duration_secs': 0.009863} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.789940] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.790221] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1167.790469] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.790621] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.790807] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1167.791091] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2998d4d6-97f3-43cb-a7e8-3a5cbb6f261d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.881557] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.714s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.883973] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.488s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.885549] env[62521]: INFO nova.compute.claims [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1167.902606] env[62521]: INFO nova.scheduler.client.report [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Deleted allocations for instance 9eeeef50-4538-4b45-b14b-46671356adbf [ 1167.987365] env[62521]: DEBUG oslo_concurrency.lockutils [req-01345df8-a2e0-478c-a413-17daa5ce8179 req-41815653-f3fa-416d-bba1-2fdc76021e95 service nova] Releasing lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.387414] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "eeb244ad-07b5-45da-808a-e50968a5dd1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.387649] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "eeb244ad-07b5-45da-808a-e50968a5dd1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.409935] env[62521]: DEBUG oslo_concurrency.lockutils [None req-0e474a15-069b-44f3-8650-7ff3b61b64c4 tempest-DeleteServersTestJSON-1301600404 tempest-DeleteServersTestJSON-1301600404-project-member] Lock "9eeeef50-4538-4b45-b14b-46671356adbf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.068s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.889916] env[62521]: DEBUG nova.compute.manager [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1169.001531] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4725d8b-c673-4bb7-a263-10a634d3a06e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.010095] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ac3ab9-e494-453f-b86e-402e6a3a007c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.039788] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8035e282-0cf2-44d9-ac7d-6919fed04d42 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.047801] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a2ef55-8b60-4188-b862-df038537a931 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.062953] env[62521]: DEBUG nova.compute.provider_tree [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.197707] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1169.197707] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1169.198567] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-945efc90-64e0-4e81-81d7-ab354bd89ddb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.205632] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1169.205632] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52aeddf3-a1c0-f402-4297-e06d74a6c6e5" [ 1169.205632] env[62521]: _type = "Task" [ 1169.205632] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.213849] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52aeddf3-a1c0-f402-4297-e06d74a6c6e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.409529] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.566457] env[62521]: DEBUG nova.scheduler.client.report [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1169.716574] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52aeddf3-a1c0-f402-4297-e06d74a6c6e5, 'name': SearchDatastore_Task, 'duration_secs': 0.047567} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.717402] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb744a4b-5576-44e4-9d34-cfb6e9d39821 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.722386] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1169.722386] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a18a0b-8da1-7365-b1ed-8ca56993a904" [ 1169.722386] env[62521]: _type = "Task" [ 1169.722386] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.729610] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a18a0b-8da1-7365-b1ed-8ca56993a904, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.071901] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.188s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.072689] env[62521]: DEBUG nova.compute.manager [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1170.076513] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.667s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.078527] env[62521]: INFO nova.compute.claims [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1170.233285] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52a18a0b-8da1-7365-b1ed-8ca56993a904, 'name': SearchDatastore_Task, 'duration_secs': 0.009567} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.233430] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1170.233700] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 89795492-b8e1-4b99-8451-3cc7a8ec3cb2/89795492-b8e1-4b99-8451-3cc7a8ec3cb2.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1170.233945] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff4a8c7c-7d0a-4de1-968a-835683353e39 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.241238] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1170.241238] env[62521]: value = "task-1319306" [ 1170.241238] env[62521]: _type = "Task" [ 1170.241238] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.249037] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319306, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.584422] env[62521]: DEBUG nova.compute.utils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1170.588226] env[62521]: DEBUG nova.compute.manager [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1170.588537] env[62521]: DEBUG nova.network.neutron [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1170.630916] env[62521]: DEBUG nova.policy [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ce6cad0be5f455eab09b1d348268329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '961924204a584b61a4a85a761821910d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1170.751511] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319306, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.937609] env[62521]: DEBUG nova.network.neutron [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Successfully created port: d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1171.088991] env[62521]: DEBUG nova.compute.manager [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1171.222057] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf8ba96-c9ff-4bb9-85de-41b493283386 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.228856] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545c45e2-25ab-4119-aef4-e0b216b9b72c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.264632] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145efb25-9f3f-4e48-a779-8f50b56a4a07 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.272744] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319306, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58575} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.274876] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 89795492-b8e1-4b99-8451-3cc7a8ec3cb2/89795492-b8e1-4b99-8451-3cc7a8ec3cb2.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1171.274998] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1171.275209] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c078b9e4-fe72-494d-a77c-cda1eabbf17b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.277979] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50bd487-cc58-4943-834a-9426c344d8cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.292306] env[62521]: DEBUG nova.compute.provider_tree [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1171.294414] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1171.294414] env[62521]: value = "task-1319307" [ 1171.294414] env[62521]: _type = "Task" [ 1171.294414] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.301956] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319307, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.796082] env[62521]: DEBUG nova.scheduler.client.report [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1171.807935] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319307, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059035} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.808263] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1171.809037] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e00f92-488f-4450-8aa4-3b053c8f9884 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.831685] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 89795492-b8e1-4b99-8451-3cc7a8ec3cb2/89795492-b8e1-4b99-8451-3cc7a8ec3cb2.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1171.832240] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e92a997-02cb-4ab9-a240-19b8e0dd6baa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.852531] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1171.852531] env[62521]: value = "task-1319308" [ 1171.852531] env[62521]: _type = "Task" [ 1171.852531] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.861848] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319308, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.100669] env[62521]: DEBUG nova.compute.manager [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1172.126881] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1172.127145] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1172.127333] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1172.127539] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1172.127689] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1172.127843] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1172.128065] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1172.128253] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1172.128431] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1172.128600] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1172.128778] env[62521]: DEBUG nova.virt.hardware [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1172.129670] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd3a00f-7bea-4e05-b0c6-837f9411c10c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.137829] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ca9691-6e44-4b9f-8d57-9ffc9ccd8296 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.304585] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.305553] env[62521]: DEBUG nova.compute.manager [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1172.363403] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319308, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.580370] env[62521]: DEBUG nova.compute.manager [req-10284c38-39ca-438f-97e2-2ddd5d3c249b req-52a30298-0092-43c5-a81a-adcb575cd537 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Received event network-vif-plugged-d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1172.580600] env[62521]: DEBUG oslo_concurrency.lockutils [req-10284c38-39ca-438f-97e2-2ddd5d3c249b req-52a30298-0092-43c5-a81a-adcb575cd537 service nova] Acquiring lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.580814] env[62521]: DEBUG oslo_concurrency.lockutils [req-10284c38-39ca-438f-97e2-2ddd5d3c249b req-52a30298-0092-43c5-a81a-adcb575cd537 service nova] Lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.581027] env[62521]: DEBUG oslo_concurrency.lockutils [req-10284c38-39ca-438f-97e2-2ddd5d3c249b req-52a30298-0092-43c5-a81a-adcb575cd537 service nova] Lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.581176] env[62521]: DEBUG nova.compute.manager [req-10284c38-39ca-438f-97e2-2ddd5d3c249b req-52a30298-0092-43c5-a81a-adcb575cd537 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] No waiting events found dispatching network-vif-plugged-d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1172.581384] env[62521]: WARNING nova.compute.manager [req-10284c38-39ca-438f-97e2-2ddd5d3c249b req-52a30298-0092-43c5-a81a-adcb575cd537 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Received unexpected event network-vif-plugged-d14f3c8a-fd95-4ccb-8bef-a94726c85d02 for instance with vm_state building and task_state spawning. [ 1172.753432] env[62521]: DEBUG nova.network.neutron [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Successfully updated port: d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1172.811141] env[62521]: DEBUG nova.compute.utils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1172.813131] env[62521]: DEBUG nova.compute.manager [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1172.813131] env[62521]: DEBUG nova.network.neutron [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1172.863754] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319308, 'name': ReconfigVM_Task, 'duration_secs': 0.563728} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.864055] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 89795492-b8e1-4b99-8451-3cc7a8ec3cb2/89795492-b8e1-4b99-8451-3cc7a8ec3cb2.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1172.864666] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fb940c31-59f9-4282-a246-5de431cea535 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.870881] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1172.870881] env[62521]: value = "task-1319309" [ 1172.870881] env[62521]: _type = "Task" [ 1172.870881] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.879561] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319309, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.880820] env[62521]: DEBUG nova.policy [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a0e35b0e4574ed9afc4549d30daa210', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '31c788848b0d47478564e53066e7c51a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1173.256042] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.256320] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.256372] env[62521]: DEBUG nova.network.neutron [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1173.315919] env[62521]: DEBUG nova.compute.manager [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1173.368129] env[62521]: DEBUG nova.network.neutron [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Successfully created port: 38561605-a90a-4826-b8fb-0ba01748eff0 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1173.386229] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319309, 'name': Rename_Task, 'duration_secs': 0.162262} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.386865] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1173.387276] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dcf7bbe7-1385-4c1b-b98e-3788f65ecfe8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.395787] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1173.395787] env[62521]: value = "task-1319310" [ 1173.395787] env[62521]: _type = "Task" [ 1173.395787] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.410078] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319310, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.793031] env[62521]: DEBUG nova.network.neutron [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1173.905717] env[62521]: DEBUG oslo_vmware.api [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319310, 'name': PowerOnVM_Task, 'duration_secs': 0.488316} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.906784] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1173.906784] env[62521]: INFO nova.compute.manager [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Took 9.63 seconds to spawn the instance on the hypervisor. [ 1173.906784] env[62521]: DEBUG nova.compute.manager [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1173.907415] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c00fb1f-c3e7-45f5-81f9-63accd890d19 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.152817] env[62521]: DEBUG nova.network.neutron [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updating instance_info_cache with network_info: [{"id": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "address": "fa:16:3e:e2:90:62", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14f3c8a-fd", "ovs_interfaceid": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.326809] env[62521]: DEBUG nova.compute.manager [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1174.353201] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1174.353201] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1174.353395] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1174.353437] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1174.353571] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1174.353726] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1174.353937] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1174.354132] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1174.354340] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1174.354512] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1174.354692] env[62521]: DEBUG nova.virt.hardware [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1174.355582] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86d7776-8fb6-4d53-8eb5-2d88e5f659ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.363921] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9092fd0d-8700-4c79-a63f-c02478adcb76 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.426026] env[62521]: INFO nova.compute.manager [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Took 15.20 seconds to build instance. [ 1174.607127] env[62521]: DEBUG nova.compute.manager [req-7902a7d2-9405-4395-b9f0-caa1f3127399 req-c18baf92-0522-4133-b46f-4c805120b6ec service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Received event network-changed-d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1174.607127] env[62521]: DEBUG nova.compute.manager [req-7902a7d2-9405-4395-b9f0-caa1f3127399 req-c18baf92-0522-4133-b46f-4c805120b6ec service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Refreshing instance network info cache due to event network-changed-d14f3c8a-fd95-4ccb-8bef-a94726c85d02. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1174.607127] env[62521]: DEBUG oslo_concurrency.lockutils [req-7902a7d2-9405-4395-b9f0-caa1f3127399 req-c18baf92-0522-4133-b46f-4c805120b6ec service nova] Acquiring lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1174.656345] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1174.656732] env[62521]: DEBUG nova.compute.manager [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Instance network_info: |[{"id": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "address": "fa:16:3e:e2:90:62", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14f3c8a-fd", "ovs_interfaceid": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1174.657111] env[62521]: DEBUG oslo_concurrency.lockutils [req-7902a7d2-9405-4395-b9f0-caa1f3127399 req-c18baf92-0522-4133-b46f-4c805120b6ec service nova] Acquired lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1174.657360] env[62521]: DEBUG nova.network.neutron [req-7902a7d2-9405-4395-b9f0-caa1f3127399 req-c18baf92-0522-4133-b46f-4c805120b6ec service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Refreshing network info cache for port d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1174.658643] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e2:90:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92cdccfd-4b10-4024-b724-5f22792dd4de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd14f3c8a-fd95-4ccb-8bef-a94726c85d02', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1174.666269] env[62521]: DEBUG oslo.service.loopingcall [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1174.667067] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1174.667315] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-727a3507-8394-49b1-bdb2-5584b8f5fd80 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.687556] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1174.687556] env[62521]: value = "task-1319311" [ 1174.687556] env[62521]: _type = "Task" [ 1174.687556] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.695337] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319311, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.927464] env[62521]: DEBUG oslo_concurrency.lockutils [None req-df3c3bbd-23c0-4d4c-9261-079017c4031d tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.713s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.003530] env[62521]: DEBUG nova.network.neutron [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Successfully updated port: 38561605-a90a-4826-b8fb-0ba01748eff0 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1175.199740] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319311, 'name': CreateVM_Task} progress is 25%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.286287] env[62521]: DEBUG nova.compute.manager [req-f7d2f8a4-07c2-4c11-a9ee-d0013b637549 req-8bfeb208-80f4-4b7f-8941-27e33980b12e service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received event network-changed-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1175.286530] env[62521]: DEBUG nova.compute.manager [req-f7d2f8a4-07c2-4c11-a9ee-d0013b637549 req-8bfeb208-80f4-4b7f-8941-27e33980b12e service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Refreshing instance network info cache due to event network-changed-c6359cbd-f300-4344-8a1b-bed3f0dccdf2. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1175.286787] env[62521]: DEBUG oslo_concurrency.lockutils [req-f7d2f8a4-07c2-4c11-a9ee-d0013b637549 req-8bfeb208-80f4-4b7f-8941-27e33980b12e service nova] Acquiring lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.286966] env[62521]: DEBUG oslo_concurrency.lockutils [req-f7d2f8a4-07c2-4c11-a9ee-d0013b637549 req-8bfeb208-80f4-4b7f-8941-27e33980b12e service nova] Acquired lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.287190] env[62521]: DEBUG nova.network.neutron [req-f7d2f8a4-07c2-4c11-a9ee-d0013b637549 req-8bfeb208-80f4-4b7f-8941-27e33980b12e service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Refreshing network info cache for port c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1175.413247] env[62521]: DEBUG nova.network.neutron [req-7902a7d2-9405-4395-b9f0-caa1f3127399 req-c18baf92-0522-4133-b46f-4c805120b6ec service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updated VIF entry in instance network info cache for port d14f3c8a-fd95-4ccb-8bef-a94726c85d02. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1175.413637] env[62521]: DEBUG nova.network.neutron [req-7902a7d2-9405-4395-b9f0-caa1f3127399 req-c18baf92-0522-4133-b46f-4c805120b6ec service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updating instance_info_cache with network_info: [{"id": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "address": "fa:16:3e:e2:90:62", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14f3c8a-fd", "ovs_interfaceid": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1175.505647] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.505805] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.505956] env[62521]: DEBUG nova.network.neutron [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1175.698507] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319311, 'name': CreateVM_Task, 'duration_secs': 0.645258} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.698507] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1175.699160] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.699382] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.699734] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1175.700053] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87bcaf35-a991-4223-a351-ff3bf31b91c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.704615] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1175.704615] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522db165-aed2-b154-9401-4771c5d4e59e" [ 1175.704615] env[62521]: _type = "Task" [ 1175.704615] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.711811] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522db165-aed2-b154-9401-4771c5d4e59e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.915957] env[62521]: DEBUG oslo_concurrency.lockutils [req-7902a7d2-9405-4395-b9f0-caa1f3127399 req-c18baf92-0522-4133-b46f-4c805120b6ec service nova] Releasing lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1175.993848] env[62521]: DEBUG nova.network.neutron [req-f7d2f8a4-07c2-4c11-a9ee-d0013b637549 req-8bfeb208-80f4-4b7f-8941-27e33980b12e service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updated VIF entry in instance network info cache for port c6359cbd-f300-4344-8a1b-bed3f0dccdf2. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1175.994235] env[62521]: DEBUG nova.network.neutron [req-f7d2f8a4-07c2-4c11-a9ee-d0013b637549 req-8bfeb208-80f4-4b7f-8941-27e33980b12e service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updating instance_info_cache with network_info: [{"id": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "address": "fa:16:3e:9f:8f:e8", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6359cbd-f3", "ovs_interfaceid": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.032231] env[62521]: DEBUG nova.network.neutron [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1176.155116] env[62521]: DEBUG nova.network.neutron [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Updating instance_info_cache with network_info: [{"id": "38561605-a90a-4826-b8fb-0ba01748eff0", "address": "fa:16:3e:d0:47:a6", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38561605-a9", "ovs_interfaceid": "38561605-a90a-4826-b8fb-0ba01748eff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.215039] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]522db165-aed2-b154-9401-4771c5d4e59e, 'name': SearchDatastore_Task, 'duration_secs': 0.009443} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.215363] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.215604] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1176.215836] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.215986] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.216188] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1176.216451] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ccb2afd-67fb-438f-86a6-c3ec599528de {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.224183] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1176.224366] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1176.225036] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-482d88d5-3e2d-4d2f-a630-af7d14eaddbf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.229895] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1176.229895] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5279e343-5cf0-34fd-079e-f52e93526607" [ 1176.229895] env[62521]: _type = "Task" [ 1176.229895] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.237776] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5279e343-5cf0-34fd-079e-f52e93526607, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.497676] env[62521]: DEBUG oslo_concurrency.lockutils [req-f7d2f8a4-07c2-4c11-a9ee-d0013b637549 req-8bfeb208-80f4-4b7f-8941-27e33980b12e service nova] Releasing lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.635186] env[62521]: DEBUG nova.compute.manager [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Received event network-vif-plugged-38561605-a90a-4826-b8fb-0ba01748eff0 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1176.635416] env[62521]: DEBUG oslo_concurrency.lockutils [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] Acquiring lock "eeb244ad-07b5-45da-808a-e50968a5dd1e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.635618] env[62521]: DEBUG oslo_concurrency.lockutils [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] Lock "eeb244ad-07b5-45da-808a-e50968a5dd1e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.635798] env[62521]: DEBUG oslo_concurrency.lockutils [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] Lock "eeb244ad-07b5-45da-808a-e50968a5dd1e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.635975] env[62521]: DEBUG nova.compute.manager [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] No waiting events found dispatching network-vif-plugged-38561605-a90a-4826-b8fb-0ba01748eff0 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1176.636356] env[62521]: WARNING nova.compute.manager [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Received unexpected event network-vif-plugged-38561605-a90a-4826-b8fb-0ba01748eff0 for instance with vm_state building and task_state spawning. [ 1176.636547] env[62521]: DEBUG nova.compute.manager [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Received event network-changed-38561605-a90a-4826-b8fb-0ba01748eff0 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1176.636710] env[62521]: DEBUG nova.compute.manager [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Refreshing instance network info cache due to event network-changed-38561605-a90a-4826-b8fb-0ba01748eff0. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1176.636917] env[62521]: DEBUG oslo_concurrency.lockutils [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] Acquiring lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.657647] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.657946] env[62521]: DEBUG nova.compute.manager [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Instance network_info: |[{"id": "38561605-a90a-4826-b8fb-0ba01748eff0", "address": "fa:16:3e:d0:47:a6", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38561605-a9", "ovs_interfaceid": "38561605-a90a-4826-b8fb-0ba01748eff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1176.658363] env[62521]: DEBUG oslo_concurrency.lockutils [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] Acquired lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.658617] env[62521]: DEBUG nova.network.neutron [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Refreshing network info cache for port 38561605-a90a-4826-b8fb-0ba01748eff0 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1176.661022] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:47:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '38561605-a90a-4826-b8fb-0ba01748eff0', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1176.669089] env[62521]: DEBUG oslo.service.loopingcall [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1176.670256] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1176.670455] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1263be3b-f8e5-40bb-970a-203b008c52f3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.690099] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1176.690099] env[62521]: value = "task-1319312" [ 1176.690099] env[62521]: _type = "Task" [ 1176.690099] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.698131] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319312, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.740267] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5279e343-5cf0-34fd-079e-f52e93526607, 'name': SearchDatastore_Task, 'duration_secs': 0.009249} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.741142] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c00cbc7-1141-4864-b11b-e84c9997c5dd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.746523] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1176.746523] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52271fd8-2087-196d-8d51-96717bffd05c" [ 1176.746523] env[62521]: _type = "Task" [ 1176.746523] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.756189] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52271fd8-2087-196d-8d51-96717bffd05c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.200909] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319312, 'name': CreateVM_Task, 'duration_secs': 0.334322} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.201089] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1177.201742] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.201915] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.202259] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1177.202511] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba3fff62-b5d7-475a-897d-511d3ce66c50 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.207480] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1177.207480] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523a99bb-e3f4-2629-02ff-a37267866d55" [ 1177.207480] env[62521]: _type = "Task" [ 1177.207480] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.214477] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523a99bb-e3f4-2629-02ff-a37267866d55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.257057] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52271fd8-2087-196d-8d51-96717bffd05c, 'name': SearchDatastore_Task, 'duration_secs': 0.012718} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.257057] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.257241] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3/762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1177.257463] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2bc929e8-43a8-4513-a196-43e0d3595e46 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.264050] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1177.264050] env[62521]: value = "task-1319313" [ 1177.264050] env[62521]: _type = "Task" [ 1177.264050] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.271993] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319313, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.351677] env[62521]: DEBUG nova.network.neutron [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Updated VIF entry in instance network info cache for port 38561605-a90a-4826-b8fb-0ba01748eff0. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1177.352093] env[62521]: DEBUG nova.network.neutron [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Updating instance_info_cache with network_info: [{"id": "38561605-a90a-4826-b8fb-0ba01748eff0", "address": "fa:16:3e:d0:47:a6", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38561605-a9", "ovs_interfaceid": "38561605-a90a-4826-b8fb-0ba01748eff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.717694] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]523a99bb-e3f4-2629-02ff-a37267866d55, 'name': SearchDatastore_Task, 'duration_secs': 0.009436} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.718061] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.718201] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1177.718439] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.718593] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.718775] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1177.719036] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2e998cc-0baa-4a7a-80ae-bf3ae676d4f0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.726430] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1177.726606] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1177.727300] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97b500b7-599e-4283-9b92-db66bd2650c9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.732221] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1177.732221] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5220cea7-59d1-6606-71e9-2310a3b0693c" [ 1177.732221] env[62521]: _type = "Task" [ 1177.732221] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.739508] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5220cea7-59d1-6606-71e9-2310a3b0693c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.772897] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319313, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.855042] env[62521]: DEBUG oslo_concurrency.lockutils [req-5e9a3851-695f-433f-9263-c5c36af78669 req-c874d7ac-5e7d-4578-8bab-e782a4753ce5 service nova] Releasing lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.242987] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5220cea7-59d1-6606-71e9-2310a3b0693c, 'name': SearchDatastore_Task, 'duration_secs': 0.009255} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.243786] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e707a6e9-922d-47d8-bb5f-e0bc9e1c26d0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.248897] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1178.248897] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524cad00-e2ca-d7a2-1bf3-416535050b5f" [ 1178.248897] env[62521]: _type = "Task" [ 1178.248897] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.257006] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524cad00-e2ca-d7a2-1bf3-416535050b5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.272762] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319313, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.983635} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.272968] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3/762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1178.273199] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1178.273449] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71cfdb51-4f11-4c88-a68e-ffc03e1afc98 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.279663] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1178.279663] env[62521]: value = "task-1319314" [ 1178.279663] env[62521]: _type = "Task" [ 1178.279663] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.287573] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319314, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.759631] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]524cad00-e2ca-d7a2-1bf3-416535050b5f, 'name': SearchDatastore_Task, 'duration_secs': 0.024034} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.759985] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.760186] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] eeb244ad-07b5-45da-808a-e50968a5dd1e/eeb244ad-07b5-45da-808a-e50968a5dd1e.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1178.760503] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f2090df-16cb-4171-877d-794f0dd89259 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.769116] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1178.769116] env[62521]: value = "task-1319315" [ 1178.769116] env[62521]: _type = "Task" [ 1178.769116] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.777923] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319315, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.786851] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319314, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.216286} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.787110] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1178.787829] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f8c718-0a18-4460-ba74-8518842448b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.808869] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3/762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1178.809145] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd2a9911-a8ef-4d6e-815d-3ededd0cd1c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.827049] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1178.827049] env[62521]: value = "task-1319316" [ 1178.827049] env[62521]: _type = "Task" [ 1178.827049] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.834438] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319316, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.279863] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319315, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.336935] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319316, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.781544] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319315, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.00545} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.781794] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] eeb244ad-07b5-45da-808a-e50968a5dd1e/eeb244ad-07b5-45da-808a-e50968a5dd1e.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1179.781998] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1179.782274] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51dcd77f-3b43-4e01-9bf4-164d32f6921a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.787707] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1179.787707] env[62521]: value = "task-1319317" [ 1179.787707] env[62521]: _type = "Task" [ 1179.787707] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.794778] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319317, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.837617] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319316, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.297457] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319317, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064809} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.297734] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1180.298506] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97003cf0-534b-45d2-ad22-53bc39898095 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.319460] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] eeb244ad-07b5-45da-808a-e50968a5dd1e/eeb244ad-07b5-45da-808a-e50968a5dd1e.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1180.319737] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edb05c56-6774-4efa-ba7a-ce4fea8de1dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.341028] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319316, 'name': ReconfigVM_Task, 'duration_secs': 1.017035} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.342184] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3/762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1180.342882] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1180.342882] env[62521]: value = "task-1319318" [ 1180.342882] env[62521]: _type = "Task" [ 1180.342882] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.343090] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29619275-2820-4f05-a6f6-847eaf3b309e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.352671] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319318, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.353773] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1180.353773] env[62521]: value = "task-1319319" [ 1180.353773] env[62521]: _type = "Task" [ 1180.353773] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.360501] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319319, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.855860] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319318, 'name': ReconfigVM_Task, 'duration_secs': 0.262051} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.858883] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Reconfigured VM instance instance-00000070 to attach disk [datastore1] eeb244ad-07b5-45da-808a-e50968a5dd1e/eeb244ad-07b5-45da-808a-e50968a5dd1e.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1180.859502] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4f444df-15eb-4309-8a5c-dfc08a109062 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.865869] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319319, 'name': Rename_Task, 'duration_secs': 0.240149} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.866991] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1180.867298] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1180.867298] env[62521]: value = "task-1319320" [ 1180.867298] env[62521]: _type = "Task" [ 1180.867298] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.867477] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7206048-85e4-4705-84a4-04e244cbfa49 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.876302] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319320, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.877413] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1180.877413] env[62521]: value = "task-1319321" [ 1180.877413] env[62521]: _type = "Task" [ 1180.877413] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.884054] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319321, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.379184] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319320, 'name': Rename_Task, 'duration_secs': 0.196232} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.382130] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1181.382389] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c15b33c-68f4-40a7-8089-f642bb3a46ae {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.388574] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319321, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.389759] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1181.389759] env[62521]: value = "task-1319322" [ 1181.389759] env[62521]: _type = "Task" [ 1181.389759] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.397800] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319322, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.888115] env[62521]: DEBUG oslo_vmware.api [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319321, 'name': PowerOnVM_Task, 'duration_secs': 0.842106} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.888548] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1181.888620] env[62521]: INFO nova.compute.manager [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Took 9.79 seconds to spawn the instance on the hypervisor. [ 1181.888775] env[62521]: DEBUG nova.compute.manager [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1181.889557] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8efa7a-e705-46d1-bd2f-6bb3a64f9ad3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.903065] env[62521]: DEBUG oslo_vmware.api [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319322, 'name': PowerOnVM_Task, 'duration_secs': 0.453642} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.903519] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1181.903729] env[62521]: INFO nova.compute.manager [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Took 7.58 seconds to spawn the instance on the hypervisor. [ 1181.903911] env[62521]: DEBUG nova.compute.manager [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1181.904626] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecad211e-f68f-458e-9c58-c4429672ebc8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.421486] env[62521]: INFO nova.compute.manager [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Took 15.05 seconds to build instance. [ 1182.424737] env[62521]: INFO nova.compute.manager [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Took 13.03 seconds to build instance. [ 1182.586218] env[62521]: DEBUG nova.compute.manager [req-f4a06198-e45e-4159-962a-eca483924265 req-e9a7841e-ba00-42fd-b3d5-37e17425b5b7 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Received event network-changed-38561605-a90a-4826-b8fb-0ba01748eff0 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1182.586534] env[62521]: DEBUG nova.compute.manager [req-f4a06198-e45e-4159-962a-eca483924265 req-e9a7841e-ba00-42fd-b3d5-37e17425b5b7 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Refreshing instance network info cache due to event network-changed-38561605-a90a-4826-b8fb-0ba01748eff0. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1182.586700] env[62521]: DEBUG oslo_concurrency.lockutils [req-f4a06198-e45e-4159-962a-eca483924265 req-e9a7841e-ba00-42fd-b3d5-37e17425b5b7 service nova] Acquiring lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.586848] env[62521]: DEBUG oslo_concurrency.lockutils [req-f4a06198-e45e-4159-962a-eca483924265 req-e9a7841e-ba00-42fd-b3d5-37e17425b5b7 service nova] Acquired lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.586998] env[62521]: DEBUG nova.network.neutron [req-f4a06198-e45e-4159-962a-eca483924265 req-e9a7841e-ba00-42fd-b3d5-37e17425b5b7 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Refreshing network info cache for port 38561605-a90a-4826-b8fb-0ba01748eff0 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1182.925474] env[62521]: DEBUG oslo_concurrency.lockutils [None req-e582bcd7-e2db-480c-aa9e-d96849c5b71f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.562s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.927055] env[62521]: DEBUG oslo_concurrency.lockutils [None req-d26731ec-f972-464c-989f-2ad01af29822 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "eeb244ad-07b5-45da-808a-e50968a5dd1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.539s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.125398] env[62521]: DEBUG nova.compute.manager [req-16446fd1-33f7-4b4b-b5dd-5dc89c3d6d5f req-d094290c-1617-4b15-824f-d36f24cad1a2 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received event network-changed-fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1183.125602] env[62521]: DEBUG nova.compute.manager [req-16446fd1-33f7-4b4b-b5dd-5dc89c3d6d5f req-d094290c-1617-4b15-824f-d36f24cad1a2 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing instance network info cache due to event network-changed-fe1d7288-4afc-44de-a9f7-4e98766784d4. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1183.125815] env[62521]: DEBUG oslo_concurrency.lockutils [req-16446fd1-33f7-4b4b-b5dd-5dc89c3d6d5f req-d094290c-1617-4b15-824f-d36f24cad1a2 service nova] Acquiring lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.125961] env[62521]: DEBUG oslo_concurrency.lockutils [req-16446fd1-33f7-4b4b-b5dd-5dc89c3d6d5f req-d094290c-1617-4b15-824f-d36f24cad1a2 service nova] Acquired lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.130328] env[62521]: DEBUG nova.network.neutron [req-16446fd1-33f7-4b4b-b5dd-5dc89c3d6d5f req-d094290c-1617-4b15-824f-d36f24cad1a2 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing network info cache for port fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1183.339687] env[62521]: DEBUG nova.network.neutron [req-f4a06198-e45e-4159-962a-eca483924265 req-e9a7841e-ba00-42fd-b3d5-37e17425b5b7 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Updated VIF entry in instance network info cache for port 38561605-a90a-4826-b8fb-0ba01748eff0. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1183.340065] env[62521]: DEBUG nova.network.neutron [req-f4a06198-e45e-4159-962a-eca483924265 req-e9a7841e-ba00-42fd-b3d5-37e17425b5b7 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Updating instance_info_cache with network_info: [{"id": "38561605-a90a-4826-b8fb-0ba01748eff0", "address": "fa:16:3e:d0:47:a6", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38561605-a9", "ovs_interfaceid": "38561605-a90a-4826-b8fb-0ba01748eff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.843057] env[62521]: DEBUG oslo_concurrency.lockutils [req-f4a06198-e45e-4159-962a-eca483924265 req-e9a7841e-ba00-42fd-b3d5-37e17425b5b7 service nova] Releasing lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.953170] env[62521]: DEBUG nova.network.neutron [req-16446fd1-33f7-4b4b-b5dd-5dc89c3d6d5f req-d094290c-1617-4b15-824f-d36f24cad1a2 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updated VIF entry in instance network info cache for port fe1d7288-4afc-44de-a9f7-4e98766784d4. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1183.953557] env[62521]: DEBUG nova.network.neutron [req-16446fd1-33f7-4b4b-b5dd-5dc89c3d6d5f req-d094290c-1617-4b15-824f-d36f24cad1a2 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updating instance_info_cache with network_info: [{"id": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "address": "fa:16:3e:08:82:2a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1d7288-4a", "ovs_interfaceid": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.456816] env[62521]: DEBUG oslo_concurrency.lockutils [req-16446fd1-33f7-4b4b-b5dd-5dc89c3d6d5f req-d094290c-1617-4b15-824f-d36f24cad1a2 service nova] Releasing lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.625247] env[62521]: DEBUG nova.compute.manager [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Received event network-changed-d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1184.625463] env[62521]: DEBUG nova.compute.manager [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Refreshing instance network info cache due to event network-changed-d14f3c8a-fd95-4ccb-8bef-a94726c85d02. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1184.625680] env[62521]: DEBUG oslo_concurrency.lockutils [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] Acquiring lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1184.625830] env[62521]: DEBUG oslo_concurrency.lockutils [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] Acquired lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.626056] env[62521]: DEBUG nova.network.neutron [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Refreshing network info cache for port d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1185.165011] env[62521]: DEBUG nova.compute.manager [req-6d6234dc-d26c-4e3e-920c-16de7936f7b9 req-c2cfccbf-ab6d-4d9e-8400-0615dcc3a750 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Received event network-changed-d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1185.165283] env[62521]: DEBUG nova.compute.manager [req-6d6234dc-d26c-4e3e-920c-16de7936f7b9 req-c2cfccbf-ab6d-4d9e-8400-0615dcc3a750 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Refreshing instance network info cache due to event network-changed-d14f3c8a-fd95-4ccb-8bef-a94726c85d02. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1185.165624] env[62521]: DEBUG oslo_concurrency.lockutils [req-6d6234dc-d26c-4e3e-920c-16de7936f7b9 req-c2cfccbf-ab6d-4d9e-8400-0615dcc3a750 service nova] Acquiring lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.330474] env[62521]: DEBUG nova.network.neutron [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updated VIF entry in instance network info cache for port d14f3c8a-fd95-4ccb-8bef-a94726c85d02. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1185.330857] env[62521]: DEBUG nova.network.neutron [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updating instance_info_cache with network_info: [{"id": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "address": "fa:16:3e:e2:90:62", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14f3c8a-fd", "ovs_interfaceid": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.833430] env[62521]: DEBUG oslo_concurrency.lockutils [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] Releasing lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.833737] env[62521]: DEBUG nova.compute.manager [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received event network-changed-fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1185.833912] env[62521]: DEBUG nova.compute.manager [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing instance network info cache due to event network-changed-fe1d7288-4afc-44de-a9f7-4e98766784d4. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1185.834146] env[62521]: DEBUG oslo_concurrency.lockutils [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] Acquiring lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.834295] env[62521]: DEBUG oslo_concurrency.lockutils [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] Acquired lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.834571] env[62521]: DEBUG nova.network.neutron [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing network info cache for port fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1185.836424] env[62521]: DEBUG oslo_concurrency.lockutils [req-6d6234dc-d26c-4e3e-920c-16de7936f7b9 req-c2cfccbf-ab6d-4d9e-8400-0615dcc3a750 service nova] Acquired lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.836424] env[62521]: DEBUG nova.network.neutron [req-6d6234dc-d26c-4e3e-920c-16de7936f7b9 req-c2cfccbf-ab6d-4d9e-8400-0615dcc3a750 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Refreshing network info cache for port d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1186.656524] env[62521]: DEBUG nova.network.neutron [req-6d6234dc-d26c-4e3e-920c-16de7936f7b9 req-c2cfccbf-ab6d-4d9e-8400-0615dcc3a750 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updated VIF entry in instance network info cache for port d14f3c8a-fd95-4ccb-8bef-a94726c85d02. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1186.656888] env[62521]: DEBUG nova.network.neutron [req-6d6234dc-d26c-4e3e-920c-16de7936f7b9 req-c2cfccbf-ab6d-4d9e-8400-0615dcc3a750 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updating instance_info_cache with network_info: [{"id": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "address": "fa:16:3e:e2:90:62", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14f3c8a-fd", "ovs_interfaceid": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.671094] env[62521]: DEBUG nova.network.neutron [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updated VIF entry in instance network info cache for port fe1d7288-4afc-44de-a9f7-4e98766784d4. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1186.671469] env[62521]: DEBUG nova.network.neutron [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updating instance_info_cache with network_info: [{"id": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "address": "fa:16:3e:08:82:2a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1d7288-4a", "ovs_interfaceid": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.160168] env[62521]: DEBUG oslo_concurrency.lockutils [req-6d6234dc-d26c-4e3e-920c-16de7936f7b9 req-c2cfccbf-ab6d-4d9e-8400-0615dcc3a750 service nova] Releasing lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.174054] env[62521]: DEBUG oslo_concurrency.lockutils [req-694551de-fbaf-408f-bbb0-8c30ebb8d8c5 req-df8da7f4-6f12-4e76-b569-7368a51b1e09 service nova] Releasing lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1196.554347] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "interface-144636f8-8650-4448-8fe5-aeec0ba6b69e-91cb8e20-a053-46d3-aa78-c841d978d8dd" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.554763] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-144636f8-8650-4448-8fe5-aeec0ba6b69e-91cb8e20-a053-46d3-aa78-c841d978d8dd" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.554992] env[62521]: DEBUG nova.objects.instance [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'flavor' on Instance uuid 144636f8-8650-4448-8fe5-aeec0ba6b69e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.152412] env[62521]: DEBUG nova.objects.instance [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'pci_requests' on Instance uuid 144636f8-8650-4448-8fe5-aeec0ba6b69e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.654861] env[62521]: DEBUG nova.objects.base [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Object Instance<144636f8-8650-4448-8fe5-aeec0ba6b69e> lazy-loaded attributes: flavor,pci_requests {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1197.655235] env[62521]: DEBUG nova.network.neutron [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1197.717473] env[62521]: DEBUG nova.policy [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ce6cad0be5f455eab09b1d348268329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '961924204a584b61a4a85a761821910d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1198.834895] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.835196] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.094204] env[62521]: DEBUG nova.compute.manager [req-cef41f32-43b5-4dbf-ad0e-5eb67d844b4f req-2e13e25a-952a-4be0-87cd-7bc08963a8a7 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received event network-vif-plugged-91cb8e20-a053-46d3-aa78-c841d978d8dd {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1199.094507] env[62521]: DEBUG oslo_concurrency.lockutils [req-cef41f32-43b5-4dbf-ad0e-5eb67d844b4f req-2e13e25a-952a-4be0-87cd-7bc08963a8a7 service nova] Acquiring lock "144636f8-8650-4448-8fe5-aeec0ba6b69e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.094647] env[62521]: DEBUG oslo_concurrency.lockutils [req-cef41f32-43b5-4dbf-ad0e-5eb67d844b4f req-2e13e25a-952a-4be0-87cd-7bc08963a8a7 service nova] Lock "144636f8-8650-4448-8fe5-aeec0ba6b69e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.094919] env[62521]: DEBUG oslo_concurrency.lockutils [req-cef41f32-43b5-4dbf-ad0e-5eb67d844b4f req-2e13e25a-952a-4be0-87cd-7bc08963a8a7 service nova] Lock "144636f8-8650-4448-8fe5-aeec0ba6b69e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.094973] env[62521]: DEBUG nova.compute.manager [req-cef41f32-43b5-4dbf-ad0e-5eb67d844b4f req-2e13e25a-952a-4be0-87cd-7bc08963a8a7 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] No waiting events found dispatching network-vif-plugged-91cb8e20-a053-46d3-aa78-c841d978d8dd {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1199.095160] env[62521]: WARNING nova.compute.manager [req-cef41f32-43b5-4dbf-ad0e-5eb67d844b4f req-2e13e25a-952a-4be0-87cd-7bc08963a8a7 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received unexpected event network-vif-plugged-91cb8e20-a053-46d3-aa78-c841d978d8dd for instance with vm_state active and task_state None. [ 1199.183268] env[62521]: DEBUG nova.network.neutron [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Successfully updated port: 91cb8e20-a053-46d3-aa78-c841d978d8dd {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1199.338491] env[62521]: INFO nova.compute.manager [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Detaching volume 1c92dcd2-2187-4e77-aac7-70adfbbfd03d [ 1199.368241] env[62521]: INFO nova.virt.block_device [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Attempting to driver detach volume 1c92dcd2-2187-4e77-aac7-70adfbbfd03d from mountpoint /dev/sdb [ 1199.368511] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Volume detach. Driver type: vmdk {{(pid=62521) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1199.368707] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282202', 'volume_id': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'name': 'volume-1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1bea570f-f2b2-4e9a-8dbb-418025402d3d', 'attached_at': '', 'detached_at': '', 'volume_id': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'serial': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1199.369613] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b6da6a-8b28-48f8-a4cc-1b9ea08ca164 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.391470] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6462d84-20ff-442d-a77a-be552f4b2803 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.398140] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564ec52a-601d-428d-a889-e78ad17fb535 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.417499] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f48e86-558d-4a50-9932-e4e80149fea1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.433437] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] The volume has not been displaced from its original location: [datastore1] volume-1c92dcd2-2187-4e77-aac7-70adfbbfd03d/volume-1c92dcd2-2187-4e77-aac7-70adfbbfd03d.vmdk. No consolidation needed. {{(pid=62521) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1199.439028] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1199.439309] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d375fab8-9070-40b6-83b5-705ca88b7c37 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.456766] env[62521]: DEBUG oslo_vmware.api [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1199.456766] env[62521]: value = "task-1319323" [ 1199.456766] env[62521]: _type = "Task" [ 1199.456766] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.464142] env[62521]: DEBUG oslo_vmware.api [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319323, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.686092] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.686338] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.686503] env[62521]: DEBUG nova.network.neutron [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1199.969540] env[62521]: DEBUG oslo_vmware.api [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319323, 'name': ReconfigVM_Task, 'duration_secs': 0.217543} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.969540] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62521) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1199.972396] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-185ea5fb-a761-4249-adcf-7a04294b0b18 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.987311] env[62521]: DEBUG oslo_vmware.api [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1199.987311] env[62521]: value = "task-1319324" [ 1199.987311] env[62521]: _type = "Task" [ 1199.987311] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.996264] env[62521]: DEBUG oslo_vmware.api [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319324, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.221919] env[62521]: WARNING nova.network.neutron [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] 67bdad13-8ff4-4bf8-b255-d0a285416fbe already exists in list: networks containing: ['67bdad13-8ff4-4bf8-b255-d0a285416fbe']. ignoring it [ 1200.497363] env[62521]: DEBUG oslo_vmware.api [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319324, 'name': ReconfigVM_Task, 'duration_secs': 0.132876} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.497633] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-282202', 'volume_id': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'name': 'volume-1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1bea570f-f2b2-4e9a-8dbb-418025402d3d', 'attached_at': '', 'detached_at': '', 'volume_id': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d', 'serial': '1c92dcd2-2187-4e77-aac7-70adfbbfd03d'} {{(pid=62521) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1200.699085] env[62521]: DEBUG nova.network.neutron [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updating instance_info_cache with network_info: [{"id": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "address": "fa:16:3e:08:82:2a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1d7288-4a", "ovs_interfaceid": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "91cb8e20-a053-46d3-aa78-c841d978d8dd", "address": "fa:16:3e:75:9a:aa", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91cb8e20-a0", "ovs_interfaceid": "91cb8e20-a053-46d3-aa78-c841d978d8dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1201.041602] env[62521]: DEBUG nova.objects.instance [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lazy-loading 'flavor' on Instance uuid 1bea570f-f2b2-4e9a-8dbb-418025402d3d {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.118707] env[62521]: DEBUG nova.compute.manager [req-d3e4cb1b-83aa-4834-979f-89bfc6ecf1d1 req-93b661cd-496a-4078-af41-daab2ba5b2c7 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received event network-changed-91cb8e20-a053-46d3-aa78-c841d978d8dd {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1201.118921] env[62521]: DEBUG nova.compute.manager [req-d3e4cb1b-83aa-4834-979f-89bfc6ecf1d1 req-93b661cd-496a-4078-af41-daab2ba5b2c7 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing instance network info cache due to event network-changed-91cb8e20-a053-46d3-aa78-c841d978d8dd. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1201.119134] env[62521]: DEBUG oslo_concurrency.lockutils [req-d3e4cb1b-83aa-4834-979f-89bfc6ecf1d1 req-93b661cd-496a-4078-af41-daab2ba5b2c7 service nova] Acquiring lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1201.201898] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.202625] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1201.202795] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.203103] env[62521]: DEBUG oslo_concurrency.lockutils [req-d3e4cb1b-83aa-4834-979f-89bfc6ecf1d1 req-93b661cd-496a-4078-af41-daab2ba5b2c7 service nova] Acquired lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.203293] env[62521]: DEBUG nova.network.neutron [req-d3e4cb1b-83aa-4834-979f-89bfc6ecf1d1 req-93b661cd-496a-4078-af41-daab2ba5b2c7 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing network info cache for port 91cb8e20-a053-46d3-aa78-c841d978d8dd {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1201.205013] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89eb2b22-1344-4a58-b27d-2cb4dcdebe47 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.226581] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1201.226816] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1201.226972] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1201.227179] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1201.227333] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1201.227489] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1201.227700] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1201.227868] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1201.228049] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1201.228221] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1201.228396] env[62521]: DEBUG nova.virt.hardware [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1201.235160] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Reconfiguring VM to attach interface {{(pid=62521) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1201.236190] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ac32ec4-2354-4a52-83f7-0a4ec7a7b6fc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.254808] env[62521]: DEBUG oslo_vmware.api [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1201.254808] env[62521]: value = "task-1319325" [ 1201.254808] env[62521]: _type = "Task" [ 1201.254808] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.262845] env[62521]: DEBUG oslo_vmware.api [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319325, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.766163] env[62521]: DEBUG oslo_vmware.api [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319325, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.905574] env[62521]: DEBUG nova.network.neutron [req-d3e4cb1b-83aa-4834-979f-89bfc6ecf1d1 req-93b661cd-496a-4078-af41-daab2ba5b2c7 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updated VIF entry in instance network info cache for port 91cb8e20-a053-46d3-aa78-c841d978d8dd. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1201.906022] env[62521]: DEBUG nova.network.neutron [req-d3e4cb1b-83aa-4834-979f-89bfc6ecf1d1 req-93b661cd-496a-4078-af41-daab2ba5b2c7 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updating instance_info_cache with network_info: [{"id": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "address": "fa:16:3e:08:82:2a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1d7288-4a", "ovs_interfaceid": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "91cb8e20-a053-46d3-aa78-c841d978d8dd", "address": "fa:16:3e:75:9a:aa", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91cb8e20-a0", "ovs_interfaceid": "91cb8e20-a053-46d3-aa78-c841d978d8dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1202.050119] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a36d6c0c-3060-4ca2-8004-df92f5afd125 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.215s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.264900] env[62521]: DEBUG oslo_vmware.api [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319325, 'name': ReconfigVM_Task, 'duration_secs': 0.535772} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.265434] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1202.265661] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Reconfigured VM to attach interface {{(pid=62521) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1202.409712] env[62521]: DEBUG oslo_concurrency.lockutils [req-d3e4cb1b-83aa-4834-979f-89bfc6ecf1d1 req-93b661cd-496a-4078-af41-daab2ba5b2c7 service nova] Releasing lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1202.770057] env[62521]: DEBUG oslo_concurrency.lockutils [None req-99a5d790-3497-4712-a02d-69b536ba841f tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-144636f8-8650-4448-8fe5-aeec0ba6b69e-91cb8e20-a053-46d3-aa78-c841d978d8dd" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.215s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.135405] env[62521]: DEBUG oslo_concurrency.lockutils [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.135764] env[62521]: DEBUG oslo_concurrency.lockutils [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.135906] env[62521]: DEBUG oslo_concurrency.lockutils [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.136113] env[62521]: DEBUG oslo_concurrency.lockutils [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.136291] env[62521]: DEBUG oslo_concurrency.lockutils [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.138453] env[62521]: INFO nova.compute.manager [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Terminating instance [ 1203.140197] env[62521]: DEBUG nova.compute.manager [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1203.140395] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1203.141254] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e0a568-da99-45a5-bdf3-5ff21e79ba7b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.149746] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1203.149909] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-871fa9b1-ad1f-45a8-96d6-b5c3fcb89411 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.156612] env[62521]: DEBUG oslo_vmware.api [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1203.156612] env[62521]: value = "task-1319326" [ 1203.156612] env[62521]: _type = "Task" [ 1203.156612] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.163998] env[62521]: DEBUG oslo_vmware.api [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319326, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.667293] env[62521]: DEBUG oslo_vmware.api [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319326, 'name': PowerOffVM_Task, 'duration_secs': 0.168058} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.667522] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1203.667741] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1203.667944] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a19385f3-b2ba-4ca9-8181-21cae00fcff3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.728247] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1203.728508] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1203.728688] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Deleting the datastore file [datastore2] 1bea570f-f2b2-4e9a-8dbb-418025402d3d {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1203.729444] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1446cc1-8d93-4612-b943-5ca55f98ce95 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.737894] env[62521]: DEBUG oslo_vmware.api [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1203.737894] env[62521]: value = "task-1319328" [ 1203.737894] env[62521]: _type = "Task" [ 1203.737894] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.745540] env[62521]: DEBUG oslo_vmware.api [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319328, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.222605] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "interface-144636f8-8650-4448-8fe5-aeec0ba6b69e-91cb8e20-a053-46d3-aa78-c841d978d8dd" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.222889] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-144636f8-8650-4448-8fe5-aeec0ba6b69e-91cb8e20-a053-46d3-aa78-c841d978d8dd" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.247274] env[62521]: DEBUG oslo_vmware.api [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319328, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153941} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.247524] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1204.247712] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1204.247893] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1204.248086] env[62521]: INFO nova.compute.manager [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1204.248325] env[62521]: DEBUG oslo.service.loopingcall [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1204.248507] env[62521]: DEBUG nova.compute.manager [-] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1204.248603] env[62521]: DEBUG nova.network.neutron [-] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1204.652462] env[62521]: DEBUG nova.compute.manager [req-64732fd8-cb8b-466f-89f8-948812a21ad8 req-e5bd8bb0-7e1a-48b5-aad8-f3813bb2722a service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Received event network-vif-deleted-02aabc96-04f0-4b09-bcf6-44443d6542af {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1204.652681] env[62521]: INFO nova.compute.manager [req-64732fd8-cb8b-466f-89f8-948812a21ad8 req-e5bd8bb0-7e1a-48b5-aad8-f3813bb2722a service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Neutron deleted interface 02aabc96-04f0-4b09-bcf6-44443d6542af; detaching it from the instance and deleting it from the info cache [ 1204.652858] env[62521]: DEBUG nova.network.neutron [req-64732fd8-cb8b-466f-89f8-948812a21ad8 req-e5bd8bb0-7e1a-48b5-aad8-f3813bb2722a service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1204.725542] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1204.725741] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1204.726663] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30b5391-73f0-4337-9307-f3f1de2202ca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.754497] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0397e3c-c890-4a93-abd0-17ee04e9af4a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.798030] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Reconfiguring VM to detach interface {{(pid=62521) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1204.799085] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-649470eb-b27b-42cb-b890-1cf92ce54c26 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.828870] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1204.828870] env[62521]: value = "task-1319329" [ 1204.828870] env[62521]: _type = "Task" [ 1204.828870] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.841102] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.128250] env[62521]: DEBUG nova.network.neutron [-] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.155287] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10c37394-6297-4ed1-95b7-cabde83af310 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.165141] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d669d49-52dc-4329-8bbe-aa2e4a42faa0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.194162] env[62521]: DEBUG nova.compute.manager [req-64732fd8-cb8b-466f-89f8-948812a21ad8 req-e5bd8bb0-7e1a-48b5-aad8-f3813bb2722a service nova] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Detach interface failed, port_id=02aabc96-04f0-4b09-bcf6-44443d6542af, reason: Instance 1bea570f-f2b2-4e9a-8dbb-418025402d3d could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1205.338680] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.631210] env[62521]: INFO nova.compute.manager [-] [instance: 1bea570f-f2b2-4e9a-8dbb-418025402d3d] Took 1.38 seconds to deallocate network for instance. [ 1205.839441] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.137406] env[62521]: DEBUG oslo_concurrency.lockutils [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.137756] env[62521]: DEBUG oslo_concurrency.lockutils [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.138014] env[62521]: DEBUG nova.objects.instance [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lazy-loading 'resources' on Instance uuid 1bea570f-f2b2-4e9a-8dbb-418025402d3d {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1206.340784] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.728895] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b1e071-e320-4007-b335-ccfcdbe6bc65 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.736501] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1337f7-2c96-42ef-bfe1-535c72ab1b09 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.766063] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ff6d74-60b5-4eb9-b109-6958aa3e9772 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.772733] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdb35f5-98a5-4dc5-8ebd-46fcd7f2a2da {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.786732] env[62521]: DEBUG nova.compute.provider_tree [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1206.839764] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.307233] env[62521]: ERROR nova.scheduler.client.report [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [req-ff5cf224-4f87-4df5-9b00-818ae06cdd37] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3d21544b-bfc8-42d0-86ca-d323b5e2628f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ff5cf224-4f87-4df5-9b00-818ae06cdd37"}]} [ 1207.324745] env[62521]: DEBUG nova.scheduler.client.report [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Refreshing inventories for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1207.341339] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.342287] env[62521]: DEBUG nova.scheduler.client.report [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updating ProviderTree inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1207.342491] env[62521]: DEBUG nova.compute.provider_tree [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1207.354580] env[62521]: DEBUG nova.scheduler.client.report [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Refreshing aggregate associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, aggregates: None {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1207.372613] env[62521]: DEBUG nova.scheduler.client.report [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Refreshing trait associations for resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f, traits: HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62521) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1207.454434] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2909da7-385c-410d-99d8-55c421de8c88 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.461736] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aabd3f8-b83b-48b2-88ff-e6b52b838a33 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.491543] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42fa472-7836-41e2-99ee-74e35dd483c4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.499013] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1599b46a-3fe6-409a-8e17-3deabea8391d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.511947] env[62521]: DEBUG nova.compute.provider_tree [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1207.842137] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.040770] env[62521]: DEBUG nova.scheduler.client.report [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 162 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1208.041079] env[62521]: DEBUG nova.compute.provider_tree [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 162 to 163 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1208.041276] env[62521]: DEBUG nova.compute.provider_tree [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 133, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1208.344519] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.546649] env[62521]: DEBUG oslo_concurrency.lockutils [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.409s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.571220] env[62521]: INFO nova.scheduler.client.report [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Deleted allocations for instance 1bea570f-f2b2-4e9a-8dbb-418025402d3d [ 1208.842782] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.079869] env[62521]: DEBUG oslo_concurrency.lockutils [None req-18970889-5704-410b-9a10-f6a487f17058 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "1bea570f-f2b2-4e9a-8dbb-418025402d3d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.944s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.343402] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.844282] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.345731] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.624573] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.624850] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.625069] env[62521]: INFO nova.compute.manager [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Shelving [ 1210.845465] env[62521]: DEBUG oslo_vmware.api [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319329, 'name': ReconfigVM_Task, 'duration_secs': 5.75646} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.845794] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1210.845885] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Reconfigured VM to detach interface {{(pid=62521) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1211.132536] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1211.132813] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aaf1efbe-47ce-48be-984d-5450a43f210f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.141483] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1211.141483] env[62521]: value = "task-1319331" [ 1211.141483] env[62521]: _type = "Task" [ 1211.141483] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.150942] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319331, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.651386] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319331, 'name': PowerOffVM_Task, 'duration_secs': 0.208722} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.651653] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1211.652450] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0945c4d-90ac-4b7b-ba29-d76d11c10cdc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.671923] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92ab962-6d5c-45f3-acc1-f3dbecf7b47f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.140899] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.141208] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.141400] env[62521]: DEBUG nova.network.neutron [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1212.182039] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Creating Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1212.182651] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4d5f1bd3-6bba-4a57-88fa-ad971e2e0e25 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.191323] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1212.191323] env[62521]: value = "task-1319332" [ 1212.191323] env[62521]: _type = "Task" [ 1212.191323] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.198175] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319332, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.703188] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319332, 'name': CreateSnapshot_Task, 'duration_secs': 0.421625} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.703514] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Created Snapshot of the VM instance {{(pid=62521) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1212.704354] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da06b73-f080-43e4-8813-9806ac550185 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.995565] env[62521]: INFO nova.network.neutron [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Port 91cb8e20-a053-46d3-aa78-c841d978d8dd from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1212.995956] env[62521]: DEBUG nova.network.neutron [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updating instance_info_cache with network_info: [{"id": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "address": "fa:16:3e:08:82:2a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1d7288-4a", "ovs_interfaceid": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.007062] env[62521]: DEBUG nova.compute.manager [req-e28f0ab2-e029-4754-9640-8be21ed86b68 req-40d2ec3e-45cd-49b0-b81f-27745866fcd1 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received event network-changed-fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1213.007303] env[62521]: DEBUG nova.compute.manager [req-e28f0ab2-e029-4754-9640-8be21ed86b68 req-40d2ec3e-45cd-49b0-b81f-27745866fcd1 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing instance network info cache due to event network-changed-fe1d7288-4afc-44de-a9f7-4e98766784d4. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1213.007471] env[62521]: DEBUG oslo_concurrency.lockutils [req-e28f0ab2-e029-4754-9640-8be21ed86b68 req-40d2ec3e-45cd-49b0-b81f-27745866fcd1 service nova] Acquiring lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1213.221896] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Creating linked-clone VM from snapshot {{(pid=62521) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1213.223277] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-12341529-dc84-4637-81a4-a5b9354cf46f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.231821] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1213.231821] env[62521]: value = "task-1319333" [ 1213.231821] env[62521]: _type = "Task" [ 1213.231821] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.240060] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319333, 'name': CloneVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.499232] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.501639] env[62521]: DEBUG oslo_concurrency.lockutils [req-e28f0ab2-e029-4754-9640-8be21ed86b68 req-40d2ec3e-45cd-49b0-b81f-27745866fcd1 service nova] Acquired lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.502254] env[62521]: DEBUG nova.network.neutron [req-e28f0ab2-e029-4754-9640-8be21ed86b68 req-40d2ec3e-45cd-49b0-b81f-27745866fcd1 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Refreshing network info cache for port fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1213.680289] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "interface-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-91cb8e20-a053-46d3-aa78-c841d978d8dd" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.680577] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-91cb8e20-a053-46d3-aa78-c841d978d8dd" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.681016] env[62521]: DEBUG nova.objects.instance [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'flavor' on Instance uuid 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.742575] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319333, 'name': CloneVM_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.004531] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7e098fdb-1a38-4de1-a1b2-ab53ccb09836 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-144636f8-8650-4448-8fe5-aeec0ba6b69e-91cb8e20-a053-46d3-aa78-c841d978d8dd" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.781s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.244939] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319333, 'name': CloneVM_Task} progress is 95%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.263678] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.263955] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.264170] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.264363] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.264540] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.267010] env[62521]: INFO nova.compute.manager [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Terminating instance [ 1214.268982] env[62521]: DEBUG nova.compute.manager [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1214.269223] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1214.270106] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3058914-0638-4569-aec1-dffdb4dec0e5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.273902] env[62521]: DEBUG nova.network.neutron [req-e28f0ab2-e029-4754-9640-8be21ed86b68 req-40d2ec3e-45cd-49b0-b81f-27745866fcd1 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updated VIF entry in instance network info cache for port fe1d7288-4afc-44de-a9f7-4e98766784d4. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1214.274254] env[62521]: DEBUG nova.network.neutron [req-e28f0ab2-e029-4754-9640-8be21ed86b68 req-40d2ec3e-45cd-49b0-b81f-27745866fcd1 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updating instance_info_cache with network_info: [{"id": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "address": "fa:16:3e:08:82:2a", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1d7288-4a", "ovs_interfaceid": "fe1d7288-4afc-44de-a9f7-4e98766784d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1214.277748] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1214.278573] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c6d8349-3392-4488-b1fd-b17faa867111 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.285382] env[62521]: DEBUG oslo_vmware.api [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1214.285382] env[62521]: value = "task-1319334" [ 1214.285382] env[62521]: _type = "Task" [ 1214.285382] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.298943] env[62521]: DEBUG oslo_vmware.api [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.330428] env[62521]: DEBUG nova.objects.instance [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'pci_requests' on Instance uuid 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.743901] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319333, 'name': CloneVM_Task, 'duration_secs': 1.237723} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.744211] env[62521]: INFO nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Created linked-clone VM from snapshot [ 1214.744973] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6545c7-fce8-4cea-b667-89fea63118e1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.752818] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Uploading image 52ee4876-2464-4aaa-856d-801a073562b6 {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1214.779535] env[62521]: DEBUG oslo_concurrency.lockutils [req-e28f0ab2-e029-4754-9640-8be21ed86b68 req-40d2ec3e-45cd-49b0-b81f-27745866fcd1 service nova] Releasing lock "refresh_cache-144636f8-8650-4448-8fe5-aeec0ba6b69e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1214.781852] env[62521]: DEBUG oslo_vmware.rw_handles [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1214.781852] env[62521]: value = "vm-282208" [ 1214.781852] env[62521]: _type = "VirtualMachine" [ 1214.781852] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1214.782155] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d34b4eaa-a3b4-4b25-8176-73ff6cd8efc2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.790618] env[62521]: DEBUG oslo_vmware.rw_handles [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lease: (returnval){ [ 1214.790618] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b13433-93af-0ba8-8fe4-7d0a34ec52e5" [ 1214.790618] env[62521]: _type = "HttpNfcLease" [ 1214.790618] env[62521]: } obtained for exporting VM: (result){ [ 1214.790618] env[62521]: value = "vm-282208" [ 1214.790618] env[62521]: _type = "VirtualMachine" [ 1214.790618] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1214.790983] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the lease: (returnval){ [ 1214.790983] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b13433-93af-0ba8-8fe4-7d0a34ec52e5" [ 1214.790983] env[62521]: _type = "HttpNfcLease" [ 1214.790983] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1214.797233] env[62521]: DEBUG oslo_vmware.api [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319334, 'name': PowerOffVM_Task, 'duration_secs': 0.222905} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.797838] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1214.798034] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1214.798289] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64d10f5e-639e-419e-bce4-c4a3ff53ff90 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.800952] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1214.800952] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b13433-93af-0ba8-8fe4-7d0a34ec52e5" [ 1214.800952] env[62521]: _type = "HttpNfcLease" [ 1214.800952] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1214.833383] env[62521]: DEBUG nova.objects.base [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Object Instance<762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3> lazy-loaded attributes: flavor,pci_requests {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1214.833613] env[62521]: DEBUG nova.network.neutron [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1214.861947] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1214.862210] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1214.862402] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Deleting the datastore file [datastore2] 4021ce0d-69ee-4c57-bb18-fd77f339ed72 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1214.862673] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd6ea535-9d9b-4f88-b365-672091729831 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.869552] env[62521]: DEBUG oslo_vmware.api [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1214.869552] env[62521]: value = "task-1319337" [ 1214.869552] env[62521]: _type = "Task" [ 1214.869552] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.878922] env[62521]: DEBUG oslo_vmware.api [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319337, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.912854] env[62521]: DEBUG nova.policy [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ce6cad0be5f455eab09b1d348268329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '961924204a584b61a4a85a761821910d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1215.035863] env[62521]: DEBUG nova.compute.manager [req-de5ce969-70f2-432b-b01a-c6a570e867da req-8639ada9-0530-4dfc-87e3-33cec896bb47 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Received event network-changed-d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1215.036077] env[62521]: DEBUG nova.compute.manager [req-de5ce969-70f2-432b-b01a-c6a570e867da req-8639ada9-0530-4dfc-87e3-33cec896bb47 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Refreshing instance network info cache due to event network-changed-d14f3c8a-fd95-4ccb-8bef-a94726c85d02. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1215.036303] env[62521]: DEBUG oslo_concurrency.lockutils [req-de5ce969-70f2-432b-b01a-c6a570e867da req-8639ada9-0530-4dfc-87e3-33cec896bb47 service nova] Acquiring lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1215.036465] env[62521]: DEBUG oslo_concurrency.lockutils [req-de5ce969-70f2-432b-b01a-c6a570e867da req-8639ada9-0530-4dfc-87e3-33cec896bb47 service nova] Acquired lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1215.036640] env[62521]: DEBUG nova.network.neutron [req-de5ce969-70f2-432b-b01a-c6a570e867da req-8639ada9-0530-4dfc-87e3-33cec896bb47 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Refreshing network info cache for port d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1215.300397] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1215.300397] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b13433-93af-0ba8-8fe4-7d0a34ec52e5" [ 1215.300397] env[62521]: _type = "HttpNfcLease" [ 1215.300397] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1215.301150] env[62521]: DEBUG oslo_vmware.rw_handles [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1215.301150] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b13433-93af-0ba8-8fe4-7d0a34ec52e5" [ 1215.301150] env[62521]: _type = "HttpNfcLease" [ 1215.301150] env[62521]: }. {{(pid=62521) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1215.301955] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eeee249-b925-4ea3-95c0-502c1aacd713 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.309781] env[62521]: DEBUG oslo_vmware.rw_handles [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c89ca3-79dd-ad8b-64cd-bb65bf1b3dc8/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1215.309981] env[62521]: DEBUG oslo_vmware.rw_handles [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c89ca3-79dd-ad8b-64cd-bb65bf1b3dc8/disk-0.vmdk for reading. {{(pid=62521) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1215.378865] env[62521]: DEBUG oslo_vmware.api [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319337, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284106} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.379182] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1215.379337] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1215.379518] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1215.379707] env[62521]: INFO nova.compute.manager [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1215.379962] env[62521]: DEBUG oslo.service.loopingcall [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1215.380170] env[62521]: DEBUG nova.compute.manager [-] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1215.380267] env[62521]: DEBUG nova.network.neutron [-] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1215.415395] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1632ad5a-7d0b-4d9b-808e-c236bf8ad0fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.830383] env[62521]: DEBUG nova.network.neutron [req-de5ce969-70f2-432b-b01a-c6a570e867da req-8639ada9-0530-4dfc-87e3-33cec896bb47 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updated VIF entry in instance network info cache for port d14f3c8a-fd95-4ccb-8bef-a94726c85d02. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1215.830817] env[62521]: DEBUG nova.network.neutron [req-de5ce969-70f2-432b-b01a-c6a570e867da req-8639ada9-0530-4dfc-87e3-33cec896bb47 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updating instance_info_cache with network_info: [{"id": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "address": "fa:16:3e:e2:90:62", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14f3c8a-fd", "ovs_interfaceid": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.138790] env[62521]: DEBUG nova.network.neutron [-] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.334790] env[62521]: DEBUG oslo_concurrency.lockutils [req-de5ce969-70f2-432b-b01a-c6a570e867da req-8639ada9-0530-4dfc-87e3-33cec896bb47 service nova] Releasing lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1216.466888] env[62521]: DEBUG nova.network.neutron [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Successfully updated port: 91cb8e20-a053-46d3-aa78-c841d978d8dd {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1216.641434] env[62521]: INFO nova.compute.manager [-] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Took 1.26 seconds to deallocate network for instance. [ 1216.970293] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1216.970506] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.970698] env[62521]: DEBUG nova.network.neutron [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1217.064548] env[62521]: DEBUG nova.compute.manager [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] [instance: 4021ce0d-69ee-4c57-bb18-fd77f339ed72] Received event network-vif-deleted-59e6ef56-abf3-4b61-ab38-86918b4f4ce1 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1217.064905] env[62521]: DEBUG nova.compute.manager [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Received event network-vif-plugged-91cb8e20-a053-46d3-aa78-c841d978d8dd {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1217.065262] env[62521]: DEBUG oslo_concurrency.lockutils [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] Acquiring lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.065529] env[62521]: DEBUG oslo_concurrency.lockutils [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] Lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.065940] env[62521]: DEBUG oslo_concurrency.lockutils [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] Lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.066559] env[62521]: DEBUG nova.compute.manager [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] No waiting events found dispatching network-vif-plugged-91cb8e20-a053-46d3-aa78-c841d978d8dd {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1217.066559] env[62521]: WARNING nova.compute.manager [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Received unexpected event network-vif-plugged-91cb8e20-a053-46d3-aa78-c841d978d8dd for instance with vm_state active and task_state None. [ 1217.066687] env[62521]: DEBUG nova.compute.manager [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Received event network-changed-91cb8e20-a053-46d3-aa78-c841d978d8dd {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1217.066947] env[62521]: DEBUG nova.compute.manager [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Refreshing instance network info cache due to event network-changed-91cb8e20-a053-46d3-aa78-c841d978d8dd. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1217.067293] env[62521]: DEBUG oslo_concurrency.lockutils [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] Acquiring lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1217.148269] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.148269] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.148269] env[62521]: DEBUG nova.objects.instance [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lazy-loading 'resources' on Instance uuid 4021ce0d-69ee-4c57-bb18-fd77f339ed72 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1217.505031] env[62521]: WARNING nova.network.neutron [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] 67bdad13-8ff4-4bf8-b255-d0a285416fbe already exists in list: networks containing: ['67bdad13-8ff4-4bf8-b255-d0a285416fbe']. ignoring it [ 1217.741085] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed0e7f0-eca5-4998-bf08-4f3f241223b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.749788] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2cbd780-fcf8-40ff-82bd-2853188bbdeb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.783166] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97fb31f-fe2d-42eb-8721-63abea587d62 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.790656] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04dddf4-5919-4334-8f0f-a116dffffb6b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.804151] env[62521]: DEBUG nova.compute.provider_tree [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1217.820681] env[62521]: DEBUG nova.network.neutron [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updating instance_info_cache with network_info: [{"id": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "address": "fa:16:3e:e2:90:62", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14f3c8a-fd", "ovs_interfaceid": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "91cb8e20-a053-46d3-aa78-c841d978d8dd", "address": "fa:16:3e:75:9a:aa", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91cb8e20-a0", "ovs_interfaceid": "91cb8e20-a053-46d3-aa78-c841d978d8dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.323604] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.324336] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.324974] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.324974] env[62521]: DEBUG oslo_concurrency.lockutils [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] Acquired lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.324974] env[62521]: DEBUG nova.network.neutron [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Refreshing network info cache for port 91cb8e20-a053-46d3-aa78-c841d978d8dd {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1218.326802] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a09729a-4181-4565-802e-c5c34a9db91e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.344558] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1218.344834] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1218.345010] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1218.345217] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1218.345371] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1218.345551] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1218.345766] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1218.345932] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1218.346142] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1218.346316] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1218.346497] env[62521]: DEBUG nova.virt.hardware [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1218.353019] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Reconfiguring VM to attach interface {{(pid=62521) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1218.354879] env[62521]: DEBUG nova.scheduler.client.report [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updated inventory for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with generation 163 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1218.355177] env[62521]: DEBUG nova.compute.provider_tree [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updating resource provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f generation from 163 to 164 during operation: update_inventory {{(pid=62521) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1218.355392] env[62521]: DEBUG nova.compute.provider_tree [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Updating inventory in ProviderTree for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1218.358586] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b891239b-0199-46e4-b891-b9654699a417 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.373060] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.225s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.380421] env[62521]: DEBUG oslo_vmware.api [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1218.380421] env[62521]: value = "task-1319338" [ 1218.380421] env[62521]: _type = "Task" [ 1218.380421] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.389325] env[62521]: DEBUG oslo_vmware.api [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319338, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.391119] env[62521]: INFO nova.scheduler.client.report [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Deleted allocations for instance 4021ce0d-69ee-4c57-bb18-fd77f339ed72 [ 1218.891013] env[62521]: DEBUG oslo_vmware.api [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319338, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.898073] env[62521]: DEBUG oslo_concurrency.lockutils [None req-5d359c26-389e-4b5f-a9ba-bf071dcd7ebc tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "4021ce0d-69ee-4c57-bb18-fd77f339ed72" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.634s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.052040] env[62521]: DEBUG nova.network.neutron [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updated VIF entry in instance network info cache for port 91cb8e20-a053-46d3-aa78-c841d978d8dd. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1219.052549] env[62521]: DEBUG nova.network.neutron [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updating instance_info_cache with network_info: [{"id": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "address": "fa:16:3e:e2:90:62", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14f3c8a-fd", "ovs_interfaceid": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "91cb8e20-a053-46d3-aa78-c841d978d8dd", "address": "fa:16:3e:75:9a:aa", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91cb8e20-a0", "ovs_interfaceid": "91cb8e20-a053-46d3-aa78-c841d978d8dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.175318] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.175572] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.175723] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1219.391076] env[62521]: DEBUG oslo_vmware.api [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319338, 'name': ReconfigVM_Task, 'duration_secs': 0.783257} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.391654] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.391874] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Reconfigured VM to attach interface {{(pid=62521) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1219.555356] env[62521]: DEBUG oslo_concurrency.lockutils [req-4d7b8035-66eb-4ad2-bfb9-6465277faa15 req-e6042280-1cea-4a6c-a938-f03a6c017b89 service nova] Releasing lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.805950] env[62521]: DEBUG oslo_concurrency.lockutils [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.806222] env[62521]: DEBUG oslo_concurrency.lockutils [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1219.806451] env[62521]: DEBUG oslo_concurrency.lockutils [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.806647] env[62521]: DEBUG oslo_concurrency.lockutils [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1219.806836] env[62521]: DEBUG oslo_concurrency.lockutils [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.809258] env[62521]: INFO nova.compute.manager [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Terminating instance [ 1219.811332] env[62521]: DEBUG nova.compute.manager [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1219.811570] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1219.812493] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b625e23b-9736-4514-aeff-e43362e43341 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.820582] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1219.820866] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e71df7d-1a57-4d4f-8afd-a327e75a4ee4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.827055] env[62521]: DEBUG oslo_vmware.api [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1219.827055] env[62521]: value = "task-1319339" [ 1219.827055] env[62521]: _type = "Task" [ 1219.827055] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.834957] env[62521]: DEBUG oslo_vmware.api [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319339, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.896773] env[62521]: DEBUG oslo_concurrency.lockutils [None req-1127f140-9b77-40e0-a0af-898e0fc94d4a tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-91cb8e20-a053-46d3-aa78-c841d978d8dd" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.216s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.980208] env[62521]: DEBUG oslo_concurrency.lockutils [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "eeb244ad-07b5-45da-808a-e50968a5dd1e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.980600] env[62521]: DEBUG oslo_concurrency.lockutils [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "eeb244ad-07b5-45da-808a-e50968a5dd1e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1219.980878] env[62521]: DEBUG nova.compute.manager [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1219.981888] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830f0423-8289-4b33-9648-89b28622e6c2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.988770] env[62521]: DEBUG nova.compute.manager [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62521) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1219.989358] env[62521]: DEBUG nova.objects.instance [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'flavor' on Instance uuid eeb244ad-07b5-45da-808a-e50968a5dd1e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1220.336200] env[62521]: DEBUG oslo_vmware.api [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319339, 'name': PowerOffVM_Task, 'duration_secs': 0.19558} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.336476] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1220.336652] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1220.336917] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bf57995-bb28-439b-a151-d0e4aa53ef78 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.408850] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1220.409126] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1220.409278] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Deleting the datastore file [datastore2] 703ff423-98b1-4942-b6fd-8b95fe57bd0a {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1220.409559] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19000279-27f4-4014-9205-5df495e3a0cb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.416235] env[62521]: DEBUG oslo_vmware.api [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for the task: (returnval){ [ 1220.416235] env[62521]: value = "task-1319341" [ 1220.416235] env[62521]: _type = "Task" [ 1220.416235] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.423932] env[62521]: DEBUG oslo_vmware.api [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319341, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.494508] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1220.494786] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-163667e7-46bc-4b2b-a3af-8d8b34564a1d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.501104] env[62521]: DEBUG oslo_vmware.api [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1220.501104] env[62521]: value = "task-1319342" [ 1220.501104] env[62521]: _type = "Task" [ 1220.501104] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.511342] env[62521]: DEBUG oslo_vmware.api [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319342, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.928170] env[62521]: DEBUG oslo_vmware.api [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Task: {'id': task-1319341, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.327966} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.928593] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1220.928956] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1220.929294] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1220.929579] env[62521]: INFO nova.compute.manager [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1220.929959] env[62521]: DEBUG oslo.service.loopingcall [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1220.930279] env[62521]: DEBUG nova.compute.manager [-] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1220.930431] env[62521]: DEBUG nova.network.neutron [-] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1221.011018] env[62521]: DEBUG oslo_vmware.api [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319342, 'name': PowerOffVM_Task, 'duration_secs': 0.198431} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.011340] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1221.011522] env[62521]: DEBUG nova.compute.manager [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1221.012366] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b760cb4a-6286-4daf-8d71-2652fd37ea42 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.211771] env[62521]: DEBUG nova.compute.manager [req-ea109ecf-9d7e-4b44-a171-c9df9a53f9db req-23e4ae5f-d95c-474f-a8e6-df6c105d21e4 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Received event network-vif-deleted-ee2404e8-7a05-4516-826e-2f5e4adcae45 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1221.212086] env[62521]: INFO nova.compute.manager [req-ea109ecf-9d7e-4b44-a171-c9df9a53f9db req-23e4ae5f-d95c-474f-a8e6-df6c105d21e4 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Neutron deleted interface ee2404e8-7a05-4516-826e-2f5e4adcae45; detaching it from the instance and deleting it from the info cache [ 1221.212234] env[62521]: DEBUG nova.network.neutron [req-ea109ecf-9d7e-4b44-a171-c9df9a53f9db req-23e4ae5f-d95c-474f-a8e6-df6c105d21e4 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1221.369365] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "interface-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-91cb8e20-a053-46d3-aa78-c841d978d8dd" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1221.369669] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-91cb8e20-a053-46d3-aa78-c841d978d8dd" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1221.525108] env[62521]: DEBUG oslo_concurrency.lockutils [None req-529a2cc6-456c-44f4-a1c0-1830ee00ed59 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "eeb244ad-07b5-45da-808a-e50968a5dd1e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1221.679244] env[62521]: DEBUG nova.network.neutron [-] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1221.715861] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b027e873-5499-480b-9c54-7ac03bd5b583 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.725723] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47d8f49-49ac-473a-aded-3af272ab1841 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.753900] env[62521]: DEBUG nova.compute.manager [req-ea109ecf-9d7e-4b44-a171-c9df9a53f9db req-23e4ae5f-d95c-474f-a8e6-df6c105d21e4 service nova] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Detach interface failed, port_id=ee2404e8-7a05-4516-826e-2f5e4adcae45, reason: Instance 703ff423-98b1-4942-b6fd-8b95fe57bd0a could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1221.872509] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1221.872755] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.873869] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2672d0bc-769b-460c-a890-c3e5a7015e47 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.895596] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1abc9a-5cd9-4532-afd8-e7ae575ea150 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.925288] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Reconfiguring VM to detach interface {{(pid=62521) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1221.926252] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b6d5bde-4c4f-412d-9fad-25588858cda7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.947580] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1221.947580] env[62521]: value = "task-1319343" [ 1221.947580] env[62521]: _type = "Task" [ 1221.947580] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.956047] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.182571] env[62521]: INFO nova.compute.manager [-] [instance: 703ff423-98b1-4942-b6fd-8b95fe57bd0a] Took 1.25 seconds to deallocate network for instance. [ 1222.402855] env[62521]: DEBUG nova.objects.instance [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'flavor' on Instance uuid eeb244ad-07b5-45da-808a-e50968a5dd1e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.458445] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.689586] env[62521]: DEBUG oslo_concurrency.lockutils [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.689863] env[62521]: DEBUG oslo_concurrency.lockutils [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.690103] env[62521]: DEBUG nova.objects.instance [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lazy-loading 'resources' on Instance uuid 703ff423-98b1-4942-b6fd-8b95fe57bd0a {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.695092] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Didn't find any instances for network info cache update. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1222.695292] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.695451] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.695609] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.695746] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.695876] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.696033] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.696169] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1222.696329] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.908479] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1222.908677] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.908851] env[62521]: DEBUG nova.network.neutron [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1222.909070] env[62521]: DEBUG nova.objects.instance [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'info_cache' on Instance uuid eeb244ad-07b5-45da-808a-e50968a5dd1e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.959758] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.200626] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.265615] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678bf249-fec1-41de-921d-77c6defd7276 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.272966] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77038e4f-e191-421a-b509-17e96bc3817c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.302680] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe52a1d-049a-44b6-8883-943b1d2b5db5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.310588] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4eac3d-fd36-4377-b970-233026ab9085 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.323640] env[62521]: DEBUG nova.compute.provider_tree [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1223.332741] env[62521]: DEBUG oslo_vmware.rw_handles [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c89ca3-79dd-ad8b-64cd-bb65bf1b3dc8/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1223.333793] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219d6b7e-5572-4e60-8f83-735f6e697107 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.339928] env[62521]: DEBUG oslo_vmware.rw_handles [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c89ca3-79dd-ad8b-64cd-bb65bf1b3dc8/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1223.340113] env[62521]: ERROR oslo_vmware.rw_handles [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c89ca3-79dd-ad8b-64cd-bb65bf1b3dc8/disk-0.vmdk due to incomplete transfer. [ 1223.340330] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e3543ec7-c080-45c0-87b3-b08b9616fc03 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.347155] env[62521]: DEBUG oslo_vmware.rw_handles [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c89ca3-79dd-ad8b-64cd-bb65bf1b3dc8/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1223.347380] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Uploaded image 52ee4876-2464-4aaa-856d-801a073562b6 to the Glance image server {{(pid=62521) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1223.350013] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Destroying the VM {{(pid=62521) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1223.350472] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d55a823e-12c3-4832-8bf8-6108d36388fd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.357525] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1223.357525] env[62521]: value = "task-1319344" [ 1223.357525] env[62521]: _type = "Task" [ 1223.357525] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.365086] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319344, 'name': Destroy_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.412911] env[62521]: DEBUG nova.objects.base [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1223.463224] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.828048] env[62521]: DEBUG nova.scheduler.client.report [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1223.866718] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319344, 'name': Destroy_Task, 'duration_secs': 0.337657} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.866935] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Destroyed the VM [ 1223.867206] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Deleting Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1223.867458] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-402664dc-4a5e-4466-a2e1-471306b70369 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.873503] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1223.873503] env[62521]: value = "task-1319345" [ 1223.873503] env[62521]: _type = "Task" [ 1223.873503] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.880705] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319345, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.959388] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.123470] env[62521]: DEBUG nova.network.neutron [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Updating instance_info_cache with network_info: [{"id": "38561605-a90a-4826-b8fb-0ba01748eff0", "address": "fa:16:3e:d0:47:a6", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38561605-a9", "ovs_interfaceid": "38561605-a90a-4826-b8fb-0ba01748eff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.332608] env[62521]: DEBUG oslo_concurrency.lockutils [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.643s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.334979] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.134s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.335180] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.335339] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1224.336296] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23310073-1d08-4769-86c6-6cbd32660018 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.343905] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326e38fe-c071-46b6-a44b-06dd7505e487 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.357458] env[62521]: INFO nova.scheduler.client.report [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Deleted allocations for instance 703ff423-98b1-4942-b6fd-8b95fe57bd0a [ 1224.359102] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9405e46a-78b8-4351-83f5-6b9b95db6ca1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.367745] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a4a695-621c-43d8-b5de-7e34916e4239 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.396893] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180464MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1224.397057] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.397260] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.406834] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319345, 'name': RemoveSnapshot_Task, 'duration_secs': 0.428601} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.407916] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Deleted Snapshot of the VM instance {{(pid=62521) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1224.407999] env[62521]: DEBUG nova.compute.manager [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1224.408737] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7e41fe-a365-4502-8efd-a4cb02b91acf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.460413] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.626054] env[62521]: DEBUG oslo_concurrency.lockutils [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1224.866790] env[62521]: DEBUG oslo_concurrency.lockutils [None req-593a9f91-4cce-4325-ac03-5c9451bb3ad0 tempest-ServerRescueNegativeTestJSON-1743222282 tempest-ServerRescueNegativeTestJSON-1743222282-project-member] Lock "703ff423-98b1-4942-b6fd-8b95fe57bd0a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.061s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.920558] env[62521]: INFO nova.compute.manager [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Shelve offloading [ 1224.922222] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1224.922464] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5f964dd-33db-4a3a-a7ef-953302cfeee0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.929204] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1224.929204] env[62521]: value = "task-1319346" [ 1224.929204] env[62521]: _type = "Task" [ 1224.929204] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.937254] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319346, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.959713] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.128924] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1225.129316] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0006f9c4-85e7-4a0c-8ad8-76b712342135 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.137210] env[62521]: DEBUG oslo_vmware.api [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1225.137210] env[62521]: value = "task-1319347" [ 1225.137210] env[62521]: _type = "Task" [ 1225.137210] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.144823] env[62521]: DEBUG oslo_vmware.api [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.425988] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 144636f8-8650-4448-8fe5-aeec0ba6b69e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1225.426126] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1225.426244] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1225.426366] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance eeb244ad-07b5-45da-808a-e50968a5dd1e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1225.426550] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1225.426692] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1225.440351] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] VM already powered off {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1225.440351] env[62521]: DEBUG nova.compute.manager [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1225.442164] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6c3900-2d12-4a66-8c89-1e68bedf319d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.452332] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1225.452332] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.452332] env[62521]: DEBUG nova.network.neutron [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1225.465018] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.494381] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19b5d96-ca3b-4a0c-be5a-7db5a1713ae9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.502846] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e3fad77-cc9d-4776-8a24-88a4ad4649ce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.533810] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e5df49-6100-4fee-b0e6-48252909c17f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.541918] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f166954-d88b-4268-8d19-56c98f38be36 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.554614] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1225.647544] env[62521]: DEBUG oslo_vmware.api [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319347, 'name': PowerOnVM_Task, 'duration_secs': 0.390021} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.647807] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1225.648010] env[62521]: DEBUG nova.compute.manager [None req-a8216322-3fa0-4c68-8b47-8180cbde710b tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1225.648856] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df9e22c-a2fa-492d-bf89-4ad8fe9ae20f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.964252] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.060730] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1226.459217] env[62521]: DEBUG nova.network.neutron [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updating instance_info_cache with network_info: [{"id": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "address": "fa:16:3e:9f:8f:e8", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6359cbd-f3", "ovs_interfaceid": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.466209] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.566726] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1226.566726] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.169s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.964055] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1226.965308] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.314846] env[62521]: DEBUG nova.compute.manager [req-a76bd02a-d61a-4b37-ab03-46a52ddc4c53 req-e7924cd1-e97a-4ed1-8e1e-d2d51ba84bdf service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received event network-vif-unplugged-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1227.315087] env[62521]: DEBUG oslo_concurrency.lockutils [req-a76bd02a-d61a-4b37-ab03-46a52ddc4c53 req-e7924cd1-e97a-4ed1-8e1e-d2d51ba84bdf service nova] Acquiring lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.315478] env[62521]: DEBUG oslo_concurrency.lockutils [req-a76bd02a-d61a-4b37-ab03-46a52ddc4c53 req-e7924cd1-e97a-4ed1-8e1e-d2d51ba84bdf service nova] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.315659] env[62521]: DEBUG oslo_concurrency.lockutils [req-a76bd02a-d61a-4b37-ab03-46a52ddc4c53 req-e7924cd1-e97a-4ed1-8e1e-d2d51ba84bdf service nova] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.315830] env[62521]: DEBUG nova.compute.manager [req-a76bd02a-d61a-4b37-ab03-46a52ddc4c53 req-e7924cd1-e97a-4ed1-8e1e-d2d51ba84bdf service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] No waiting events found dispatching network-vif-unplugged-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1227.316017] env[62521]: WARNING nova.compute.manager [req-a76bd02a-d61a-4b37-ab03-46a52ddc4c53 req-e7924cd1-e97a-4ed1-8e1e-d2d51ba84bdf service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received unexpected event network-vif-unplugged-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 for instance with vm_state shelved and task_state shelving_offloading. [ 1227.402844] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1227.403841] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693d9b20-db48-4b58-9595-6b358a44a26f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.413260] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1227.413509] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2236756a-6f53-406c-8360-bbd0d13b144c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.418602] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d60c36-04ae-4a24-bc0c-f62c6f9089a1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.423983] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-08cee683-ead9-4378-9634-6e85da110efa tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Suspending the VM {{(pid=62521) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1227.424210] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-1d5be455-3e17-4593-88c7-d093feef431e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.429163] env[62521]: DEBUG oslo_vmware.api [None req-08cee683-ead9-4378-9634-6e85da110efa tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1227.429163] env[62521]: value = "task-1319349" [ 1227.429163] env[62521]: _type = "Task" [ 1227.429163] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.436147] env[62521]: DEBUG oslo_vmware.api [None req-08cee683-ead9-4378-9634-6e85da110efa tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319349, 'name': SuspendVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.463039] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.474253] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1227.474546] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1227.474793] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleting the datastore file [datastore1] 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1227.475132] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b3cf1e3-cd29-4eb9-bb56-9380d799b225 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.481817] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1227.481817] env[62521]: value = "task-1319350" [ 1227.481817] env[62521]: _type = "Task" [ 1227.481817] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.489779] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.939074] env[62521]: DEBUG oslo_vmware.api [None req-08cee683-ead9-4378-9634-6e85da110efa tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319349, 'name': SuspendVM_Task} progress is 66%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.962667] env[62521]: DEBUG oslo_vmware.api [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319343, 'name': ReconfigVM_Task, 'duration_secs': 5.825222} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.962931] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.963183] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Reconfigured VM to detach interface {{(pid=62521) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1227.990206] env[62521]: DEBUG oslo_vmware.api [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319350, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.294191} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.992420] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1227.992628] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1227.992834] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1228.013967] env[62521]: INFO nova.scheduler.client.report [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleted allocations for instance 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 [ 1228.439723] env[62521]: DEBUG oslo_vmware.api [None req-08cee683-ead9-4378-9634-6e85da110efa tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319349, 'name': SuspendVM_Task, 'duration_secs': 0.64692} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.440013] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-08cee683-ead9-4378-9634-6e85da110efa tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Suspended the VM {{(pid=62521) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1228.440208] env[62521]: DEBUG nova.compute.manager [None req-08cee683-ead9-4378-9634-6e85da110efa tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1228.440986] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9144d83c-a719-42c2-bd7b-82c76b436443 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.518591] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1228.518937] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1228.519118] env[62521]: DEBUG nova.objects.instance [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'resources' on Instance uuid 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1229.021807] env[62521]: DEBUG nova.objects.instance [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'numa_topology' on Instance uuid 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1229.394489] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1229.394709] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquired lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1229.394961] env[62521]: DEBUG nova.network.neutron [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1229.398196] env[62521]: DEBUG nova.compute.manager [req-628f1abb-9928-4c02-9f54-392c06428a25 req-5d7747e4-140f-4e8c-b6cf-1d309b4d2e3a service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received event network-changed-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1229.398392] env[62521]: DEBUG nova.compute.manager [req-628f1abb-9928-4c02-9f54-392c06428a25 req-5d7747e4-140f-4e8c-b6cf-1d309b4d2e3a service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Refreshing instance network info cache due to event network-changed-c6359cbd-f300-4344-8a1b-bed3f0dccdf2. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1229.398599] env[62521]: DEBUG oslo_concurrency.lockutils [req-628f1abb-9928-4c02-9f54-392c06428a25 req-5d7747e4-140f-4e8c-b6cf-1d309b4d2e3a service nova] Acquiring lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1229.398743] env[62521]: DEBUG oslo_concurrency.lockutils [req-628f1abb-9928-4c02-9f54-392c06428a25 req-5d7747e4-140f-4e8c-b6cf-1d309b4d2e3a service nova] Acquired lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1229.398903] env[62521]: DEBUG nova.network.neutron [req-628f1abb-9928-4c02-9f54-392c06428a25 req-5d7747e4-140f-4e8c-b6cf-1d309b4d2e3a service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Refreshing network info cache for port c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1229.524526] env[62521]: DEBUG nova.objects.base [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Object Instance<89795492-b8e1-4b99-8451-3cc7a8ec3cb2> lazy-loaded attributes: resources,numa_topology {{(pid=62521) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1229.585365] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3026fb-cb34-4bab-a431-18bcdc3a6004 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.592620] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f0c48c-7097-4f53-b7da-d2f0dc5db2bb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.621057] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d70189-44f4-4b4e-ab48-1510a9534689 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.627646] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7d9a6d-a120-4166-bf44-51f5aff31cb1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.641342] env[62521]: DEBUG nova.compute.provider_tree [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1229.650208] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.702540] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1229.756330] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.756632] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.756851] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.757057] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.757238] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.759258] env[62521]: INFO nova.compute.manager [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Terminating instance [ 1229.760974] env[62521]: DEBUG nova.compute.manager [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1229.761194] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1229.762049] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d764bf1c-1e0f-4494-bda0-e25751cf1544 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.769432] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1229.769665] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8009a6bc-ad36-484f-bc38-cde928b0d0ac {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.775889] env[62521]: DEBUG oslo_vmware.api [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1229.775889] env[62521]: value = "task-1319351" [ 1229.775889] env[62521]: _type = "Task" [ 1229.775889] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.782916] env[62521]: DEBUG oslo_vmware.api [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319351, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.929662] env[62521]: INFO nova.compute.manager [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Resuming [ 1229.930848] env[62521]: DEBUG nova.objects.instance [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'flavor' on Instance uuid eeb244ad-07b5-45da-808a-e50968a5dd1e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1230.128655] env[62521]: INFO nova.network.neutron [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Port 91cb8e20-a053-46d3-aa78-c841d978d8dd from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1230.129044] env[62521]: DEBUG nova.network.neutron [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updating instance_info_cache with network_info: [{"id": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "address": "fa:16:3e:e2:90:62", "network": {"id": "67bdad13-8ff4-4bf8-b255-d0a285416fbe", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1289482496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "961924204a584b61a4a85a761821910d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd14f3c8a-fd", "ovs_interfaceid": "d14f3c8a-fd95-4ccb-8bef-a94726c85d02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.130661] env[62521]: DEBUG nova.network.neutron [req-628f1abb-9928-4c02-9f54-392c06428a25 req-5d7747e4-140f-4e8c-b6cf-1d309b4d2e3a service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updated VIF entry in instance network info cache for port c6359cbd-f300-4344-8a1b-bed3f0dccdf2. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1230.130972] env[62521]: DEBUG nova.network.neutron [req-628f1abb-9928-4c02-9f54-392c06428a25 req-5d7747e4-140f-4e8c-b6cf-1d309b4d2e3a service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updating instance_info_cache with network_info: [{"id": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "address": "fa:16:3e:9f:8f:e8", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": null, "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapc6359cbd-f3", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.144415] env[62521]: DEBUG nova.scheduler.client.report [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1230.286298] env[62521]: DEBUG oslo_vmware.api [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319351, 'name': PowerOffVM_Task, 'duration_secs': 0.202808} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.286574] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1230.286747] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1230.286996] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce5cee00-7493-4695-a47c-39d20830abce {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.344348] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1230.344585] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1230.344776] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleting the datastore file [datastore1] 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1230.345048] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7213266d-ba18-4310-84b0-c84da46955b1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.351675] env[62521]: DEBUG oslo_vmware.api [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1230.351675] env[62521]: value = "task-1319353" [ 1230.351675] env[62521]: _type = "Task" [ 1230.351675] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.359115] env[62521]: DEBUG oslo_vmware.api [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319353, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.633527] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Releasing lock "refresh_cache-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1230.635876] env[62521]: DEBUG oslo_concurrency.lockutils [req-628f1abb-9928-4c02-9f54-392c06428a25 req-5d7747e4-140f-4e8c-b6cf-1d309b4d2e3a service nova] Releasing lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1230.648412] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.129s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.862097] env[62521]: DEBUG oslo_vmware.api [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319353, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141675} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.862383] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1230.862573] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1230.862753] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1230.862932] env[62521]: INFO nova.compute.manager [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1230.863247] env[62521]: DEBUG oslo.service.loopingcall [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1230.863456] env[62521]: DEBUG nova.compute.manager [-] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1230.863552] env[62521]: DEBUG nova.network.neutron [-] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1230.938751] env[62521]: DEBUG oslo_concurrency.lockutils [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.939094] env[62521]: DEBUG oslo_concurrency.lockutils [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquired lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.939329] env[62521]: DEBUG nova.network.neutron [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1231.138462] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7edef55d-73dc-48ac-83d2-459ab0a580f6 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "interface-762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3-91cb8e20-a053-46d3-aa78-c841d978d8dd" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.769s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1231.156032] env[62521]: DEBUG oslo_concurrency.lockutils [None req-04e15bfa-39a5-4df3-87fb-503b9e437fa6 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.531s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1231.156890] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.507s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.157054] env[62521]: INFO nova.compute.manager [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Unshelving [ 1231.649460] env[62521]: DEBUG nova.compute.manager [req-c6d1f9df-73c0-4578-bcf6-6b45fdb6e3f9 req-2d456c10-c0c9-41e1-a696-7ef499d05bb3 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Received event network-vif-deleted-d14f3c8a-fd95-4ccb-8bef-a94726c85d02 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1231.649759] env[62521]: INFO nova.compute.manager [req-c6d1f9df-73c0-4578-bcf6-6b45fdb6e3f9 req-2d456c10-c0c9-41e1-a696-7ef499d05bb3 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Neutron deleted interface d14f3c8a-fd95-4ccb-8bef-a94726c85d02; detaching it from the instance and deleting it from the info cache [ 1231.649852] env[62521]: DEBUG nova.network.neutron [req-c6d1f9df-73c0-4578-bcf6-6b45fdb6e3f9 req-2d456c10-c0c9-41e1-a696-7ef499d05bb3 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.687704] env[62521]: DEBUG nova.network.neutron [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Updating instance_info_cache with network_info: [{"id": "38561605-a90a-4826-b8fb-0ba01748eff0", "address": "fa:16:3e:d0:47:a6", "network": {"id": "3ab30847-4a08-4b52-9e07-85fe865d6ef7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1893022545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31c788848b0d47478564e53066e7c51a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38561605-a9", "ovs_interfaceid": "38561605-a90a-4826-b8fb-0ba01748eff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.124062] env[62521]: DEBUG nova.network.neutron [-] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.153760] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ef2d79a-c78d-42b3-9e41-436745dc3fef {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.166477] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8832850-7145-4e3d-b819-9c6e9a9f5344 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.197296] env[62521]: DEBUG oslo_concurrency.lockutils [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Releasing lock "refresh_cache-eeb244ad-07b5-45da-808a-e50968a5dd1e" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1232.197296] env[62521]: DEBUG nova.compute.manager [req-c6d1f9df-73c0-4578-bcf6-6b45fdb6e3f9 req-2d456c10-c0c9-41e1-a696-7ef499d05bb3 service nova] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Detach interface failed, port_id=d14f3c8a-fd95-4ccb-8bef-a94726c85d02, reason: Instance 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1232.198364] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.198649] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.198896] env[62521]: DEBUG nova.objects.instance [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'pci_requests' on Instance uuid 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.200575] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a80a017-d94a-42cb-b441-6eafb84cad9a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.206804] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Resuming the VM {{(pid=62521) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1232.207061] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b249f693-c03a-4f2c-ba75-8c3294d8db1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.213297] env[62521]: DEBUG oslo_vmware.api [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1232.213297] env[62521]: value = "task-1319354" [ 1232.213297] env[62521]: _type = "Task" [ 1232.213297] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.220678] env[62521]: DEBUG oslo_vmware.api [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319354, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.627540] env[62521]: INFO nova.compute.manager [-] [instance: 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3] Took 1.76 seconds to deallocate network for instance. [ 1232.706082] env[62521]: DEBUG nova.objects.instance [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'numa_topology' on Instance uuid 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.722864] env[62521]: DEBUG oslo_vmware.api [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319354, 'name': PowerOnVM_Task, 'duration_secs': 0.496657} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.723183] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Resumed the VM {{(pid=62521) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1232.723384] env[62521]: DEBUG nova.compute.manager [None req-50a98c7b-af91-4b60-85a7-304640391a8a tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1232.724185] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293378fa-6e25-4f86-a650-ea23bdeec143 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.133963] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.208872] env[62521]: INFO nova.compute.claims [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1234.276314] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a79071f-4955-4082-9abf-1a9e48a1a6d7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.283644] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9270ff7-92f4-4067-a82b-196effc5ae01 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.312210] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a11e82e-a6cd-45a7-90b6-a397ed66171e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.318885] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb5308b-4807-41ab-b990-599b34555650 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.331656] env[62521]: DEBUG nova.compute.provider_tree [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1234.590482] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "eeb244ad-07b5-45da-808a-e50968a5dd1e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.590742] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "eeb244ad-07b5-45da-808a-e50968a5dd1e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.590962] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "eeb244ad-07b5-45da-808a-e50968a5dd1e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.591260] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "eeb244ad-07b5-45da-808a-e50968a5dd1e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.591482] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "eeb244ad-07b5-45da-808a-e50968a5dd1e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.593814] env[62521]: INFO nova.compute.manager [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Terminating instance [ 1234.595546] env[62521]: DEBUG nova.compute.manager [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1234.595757] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1234.596599] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4ee276-5173-4044-b791-658a99b39098 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.604248] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1234.604471] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21c69e3c-70f6-483b-8298-aef85d1223d1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.610248] env[62521]: DEBUG oslo_vmware.api [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1234.610248] env[62521]: value = "task-1319355" [ 1234.610248] env[62521]: _type = "Task" [ 1234.610248] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.617142] env[62521]: DEBUG oslo_vmware.api [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.834700] env[62521]: DEBUG nova.scheduler.client.report [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1235.119862] env[62521]: DEBUG oslo_vmware.api [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319355, 'name': PowerOffVM_Task, 'duration_secs': 0.171278} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.120246] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1235.120432] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1235.120677] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37f1b7b1-ec6b-46c7-ab1b-1a96a88ff80c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.176960] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1235.177211] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1235.177404] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleting the datastore file [datastore1] eeb244ad-07b5-45da-808a-e50968a5dd1e {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1235.177671] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d788507c-6c01-41a1-858a-5cb768f65515 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.184198] env[62521]: DEBUG oslo_vmware.api [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for the task: (returnval){ [ 1235.184198] env[62521]: value = "task-1319357" [ 1235.184198] env[62521]: _type = "Task" [ 1235.184198] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.191769] env[62521]: DEBUG oslo_vmware.api [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319357, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.340076] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.141s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.342383] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.209s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.342623] env[62521]: DEBUG nova.objects.instance [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'resources' on Instance uuid 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1235.368847] env[62521]: INFO nova.network.neutron [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updating port c6359cbd-f300-4344-8a1b-bed3f0dccdf2 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1235.693531] env[62521]: DEBUG oslo_vmware.api [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Task: {'id': task-1319357, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136342} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.693802] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1235.693989] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1235.694186] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1235.694366] env[62521]: INFO nova.compute.manager [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1235.694603] env[62521]: DEBUG oslo.service.loopingcall [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1235.694796] env[62521]: DEBUG nova.compute.manager [-] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1235.694889] env[62521]: DEBUG nova.network.neutron [-] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1235.900684] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e953057-c779-4339-9b51-7d71c27e13f9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.907978] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87560b5-602d-4dc0-804b-2d382b85fa9c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.936955] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01d5425-e9c6-495e-8264-cd49850126bb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.943718] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75ee744-12fd-4345-a78a-e75cac12f219 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.956353] env[62521]: DEBUG nova.compute.provider_tree [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1236.127231] env[62521]: DEBUG nova.compute.manager [req-f311d79b-5d1d-488e-8d77-2e69447cb1ef req-72d967ce-3bf5-47a4-b05d-d981de8f5644 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Received event network-vif-deleted-38561605-a90a-4826-b8fb-0ba01748eff0 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1236.127495] env[62521]: INFO nova.compute.manager [req-f311d79b-5d1d-488e-8d77-2e69447cb1ef req-72d967ce-3bf5-47a4-b05d-d981de8f5644 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Neutron deleted interface 38561605-a90a-4826-b8fb-0ba01748eff0; detaching it from the instance and deleting it from the info cache [ 1236.127599] env[62521]: DEBUG nova.network.neutron [req-f311d79b-5d1d-488e-8d77-2e69447cb1ef req-72d967ce-3bf5-47a4-b05d-d981de8f5644 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1236.460064] env[62521]: DEBUG nova.scheduler.client.report [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1236.606309] env[62521]: DEBUG nova.network.neutron [-] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1236.629664] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3e697ee-7222-4437-a82a-a0a437b63e7a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.639111] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e31fe2a-6362-4a6f-b30b-78a7b4d59cca {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.663251] env[62521]: DEBUG nova.compute.manager [req-f311d79b-5d1d-488e-8d77-2e69447cb1ef req-72d967ce-3bf5-47a4-b05d-d981de8f5644 service nova] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Detach interface failed, port_id=38561605-a90a-4826-b8fb-0ba01748eff0, reason: Instance eeb244ad-07b5-45da-808a-e50968a5dd1e could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1236.793367] env[62521]: DEBUG nova.compute.manager [req-7c86870f-0d73-4a7a-a008-3c8d7f6f9cf7 req-6569c0cd-bc36-42aa-b2e7-d26e777890a1 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received event network-vif-plugged-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1236.793367] env[62521]: DEBUG oslo_concurrency.lockutils [req-7c86870f-0d73-4a7a-a008-3c8d7f6f9cf7 req-6569c0cd-bc36-42aa-b2e7-d26e777890a1 service nova] Acquiring lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1236.793367] env[62521]: DEBUG oslo_concurrency.lockutils [req-7c86870f-0d73-4a7a-a008-3c8d7f6f9cf7 req-6569c0cd-bc36-42aa-b2e7-d26e777890a1 service nova] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1236.793367] env[62521]: DEBUG oslo_concurrency.lockutils [req-7c86870f-0d73-4a7a-a008-3c8d7f6f9cf7 req-6569c0cd-bc36-42aa-b2e7-d26e777890a1 service nova] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.793367] env[62521]: DEBUG nova.compute.manager [req-7c86870f-0d73-4a7a-a008-3c8d7f6f9cf7 req-6569c0cd-bc36-42aa-b2e7-d26e777890a1 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] No waiting events found dispatching network-vif-plugged-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1236.793367] env[62521]: WARNING nova.compute.manager [req-7c86870f-0d73-4a7a-a008-3c8d7f6f9cf7 req-6569c0cd-bc36-42aa-b2e7-d26e777890a1 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received unexpected event network-vif-plugged-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 for instance with vm_state shelved_offloaded and task_state spawning. [ 1236.885705] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1236.885868] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1236.886072] env[62521]: DEBUG nova.network.neutron [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1236.965135] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.623s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.993784] env[62521]: INFO nova.scheduler.client.report [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleted allocations for instance 762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3 [ 1237.108559] env[62521]: INFO nova.compute.manager [-] [instance: eeb244ad-07b5-45da-808a-e50968a5dd1e] Took 1.41 seconds to deallocate network for instance. [ 1237.500879] env[62521]: DEBUG oslo_concurrency.lockutils [None req-cda0feec-b5ca-4ef4-9f61-c4cce5304ae0 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "762b41ee-a4ae-4c0f-a16b-dc14dc6aaff3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.744s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1237.614969] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1237.615244] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.615465] env[62521]: DEBUG nova.objects.instance [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lazy-loading 'resources' on Instance uuid eeb244ad-07b5-45da-808a-e50968a5dd1e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1237.634850] env[62521]: DEBUG nova.network.neutron [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updating instance_info_cache with network_info: [{"id": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "address": "fa:16:3e:9f:8f:e8", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6359cbd-f3", "ovs_interfaceid": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1238.138040] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1238.169513] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b5d296ab3c821e6dc573b03db704f17d',container_format='bare',created_at=2024-10-20T00:41:33Z,direct_url=,disk_format='vmdk',id=52ee4876-2464-4aaa-856d-801a073562b6,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1580168291-shelved',owner='11829634203c4967bbcff90eb904097c',properties=ImageMetaProps,protected=,size=31662080,status='active',tags=,updated_at=2024-10-20T00:41:47Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1238.169777] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1238.169936] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1238.170142] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1238.170309] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1238.170465] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1238.170678] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1238.170842] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1238.171017] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1238.171232] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1238.171420] env[62521]: DEBUG nova.virt.hardware [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1238.172328] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1886c8cd-be40-4f7f-8407-4727b9225b65 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.176451] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6693c0ab-3d14-4d72-ac10-f6aab9afe8dc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.182246] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b7ce9f-5aa4-43ee-838f-b45754b2c6bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.188216] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d384e91b-a3c8-4dcc-9906-3a66959f93a7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.198516] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:8f:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6359cbd-f300-4344-8a1b-bed3f0dccdf2', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1238.205769] env[62521]: DEBUG oslo.service.loopingcall [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1238.206306] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1238.206511] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53bb2dcd-dd21-4be0-a9e2-536670849b34 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.243597] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d899d2-3c0c-4a7c-bd53-c4c0a6f41dd8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.247787] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1238.247787] env[62521]: value = "task-1319358" [ 1238.247787] env[62521]: _type = "Task" [ 1238.247787] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.253065] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4654126d-4203-4bae-8761-ce991103203c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.261468] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319358, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.269889] env[62521]: DEBUG nova.compute.provider_tree [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1238.464142] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.464450] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.464668] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "144636f8-8650-4448-8fe5-aeec0ba6b69e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.464856] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "144636f8-8650-4448-8fe5-aeec0ba6b69e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.465049] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "144636f8-8650-4448-8fe5-aeec0ba6b69e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.467363] env[62521]: INFO nova.compute.manager [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Terminating instance [ 1238.469235] env[62521]: DEBUG nova.compute.manager [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1238.469447] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1238.470296] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71080e2-f0c9-41af-95af-e932c1fc3565 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.477964] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1238.478221] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f258b350-d632-477f-ad9e-33f6a994213b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.487904] env[62521]: DEBUG oslo_vmware.api [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1238.487904] env[62521]: value = "task-1319359" [ 1238.487904] env[62521]: _type = "Task" [ 1238.487904] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.496057] env[62521]: DEBUG oslo_vmware.api [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319359, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.758932] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319358, 'name': CreateVM_Task, 'duration_secs': 0.300859} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.759291] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1238.759776] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1238.759950] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "[datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.760372] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1238.760633] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c205d5a-94d9-4f6e-8078-4656a8a5ece3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.765028] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1238.765028] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5241ed35-96ab-02c4-6063-e1e6b118d168" [ 1238.765028] env[62521]: _type = "Task" [ 1238.765028] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.772360] env[62521]: DEBUG nova.scheduler.client.report [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1238.775247] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5241ed35-96ab-02c4-6063-e1e6b118d168, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.816222] env[62521]: DEBUG nova.compute.manager [req-15513a12-aa1e-478e-9bac-46c647311572 req-c583b1a5-0eed-48a4-b28f-7cb1139a2a89 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received event network-changed-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1238.816271] env[62521]: DEBUG nova.compute.manager [req-15513a12-aa1e-478e-9bac-46c647311572 req-c583b1a5-0eed-48a4-b28f-7cb1139a2a89 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Refreshing instance network info cache due to event network-changed-c6359cbd-f300-4344-8a1b-bed3f0dccdf2. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1238.816567] env[62521]: DEBUG oslo_concurrency.lockutils [req-15513a12-aa1e-478e-9bac-46c647311572 req-c583b1a5-0eed-48a4-b28f-7cb1139a2a89 service nova] Acquiring lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1238.816742] env[62521]: DEBUG oslo_concurrency.lockutils [req-15513a12-aa1e-478e-9bac-46c647311572 req-c583b1a5-0eed-48a4-b28f-7cb1139a2a89 service nova] Acquired lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.816912] env[62521]: DEBUG nova.network.neutron [req-15513a12-aa1e-478e-9bac-46c647311572 req-c583b1a5-0eed-48a4-b28f-7cb1139a2a89 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Refreshing network info cache for port c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1238.997622] env[62521]: DEBUG oslo_vmware.api [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319359, 'name': PowerOffVM_Task, 'duration_secs': 0.190163} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.997832] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1238.998113] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1238.998354] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-690e0859-b862-4e62-a7b6-ba3ca3db5fec {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.056343] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1239.056671] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1239.056941] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleting the datastore file [datastore1] 144636f8-8650-4448-8fe5-aeec0ba6b69e {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1239.057291] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6eea9812-0897-468a-a357-7828eb94fa99 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.063644] env[62521]: DEBUG oslo_vmware.api [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for the task: (returnval){ [ 1239.063644] env[62521]: value = "task-1319361" [ 1239.063644] env[62521]: _type = "Task" [ 1239.063644] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.073325] env[62521]: DEBUG oslo_vmware.api [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319361, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.274730] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "[datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1239.274959] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Processing image 52ee4876-2464-4aaa-856d-801a073562b6 {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1239.275313] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6/52ee4876-2464-4aaa-856d-801a073562b6.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1239.275505] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquired lock "[datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6/52ee4876-2464-4aaa-856d-801a073562b6.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.275719] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1239.276473] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.661s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.278343] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ee056fb-2d9f-49ee-9ed0-d466892cd962 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.286414] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1239.286684] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1239.287442] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ae95eb2-ef5f-48c5-a8b8-f513c33f50df {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.292574] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1239.292574] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]528cf7cc-6706-874e-28f4-ae7e51920e77" [ 1239.292574] env[62521]: _type = "Task" [ 1239.292574] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.299825] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]528cf7cc-6706-874e-28f4-ae7e51920e77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.300703] env[62521]: INFO nova.scheduler.client.report [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Deleted allocations for instance eeb244ad-07b5-45da-808a-e50968a5dd1e [ 1239.521722] env[62521]: DEBUG nova.network.neutron [req-15513a12-aa1e-478e-9bac-46c647311572 req-c583b1a5-0eed-48a4-b28f-7cb1139a2a89 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updated VIF entry in instance network info cache for port c6359cbd-f300-4344-8a1b-bed3f0dccdf2. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1239.522050] env[62521]: DEBUG nova.network.neutron [req-15513a12-aa1e-478e-9bac-46c647311572 req-c583b1a5-0eed-48a4-b28f-7cb1139a2a89 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updating instance_info_cache with network_info: [{"id": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "address": "fa:16:3e:9f:8f:e8", "network": {"id": "c88c5ba5-88e5-4295-a8f3-65aead37dee6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2125339901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11829634203c4967bbcff90eb904097c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6359cbd-f3", "ovs_interfaceid": "c6359cbd-f300-4344-8a1b-bed3f0dccdf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1239.574672] env[62521]: DEBUG oslo_vmware.api [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Task: {'id': task-1319361, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146921} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.574939] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1239.575149] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1239.575331] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1239.575507] env[62521]: INFO nova.compute.manager [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1239.575760] env[62521]: DEBUG oslo.service.loopingcall [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1239.575930] env[62521]: DEBUG nova.compute.manager [-] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1239.576062] env[62521]: DEBUG nova.network.neutron [-] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1239.810682] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Preparing fetch location {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1239.810682] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Fetch image to [datastore2] OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125/OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125.vmdk {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1239.810967] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Downloading stream optimized image 52ee4876-2464-4aaa-856d-801a073562b6 to [datastore2] OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125/OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125.vmdk on the data store datastore2 as vApp {{(pid=62521) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1239.811163] env[62521]: DEBUG nova.virt.vmwareapi.images [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Downloading image file data 52ee4876-2464-4aaa-856d-801a073562b6 to the ESX as VM named 'OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125' {{(pid=62521) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1239.814049] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6b4a0f0d-ff1e-4527-83a7-efc47c2a27f5 tempest-ServerActionsTestJSON-434372076 tempest-ServerActionsTestJSON-434372076-project-member] Lock "eeb244ad-07b5-45da-808a-e50968a5dd1e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.223s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.887540] env[62521]: DEBUG oslo_vmware.rw_handles [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1239.887540] env[62521]: value = "resgroup-9" [ 1239.887540] env[62521]: _type = "ResourcePool" [ 1239.887540] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1239.887883] env[62521]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-1630b86d-f68e-4e2c-856e-0864d8219f1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.908909] env[62521]: DEBUG oslo_vmware.rw_handles [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lease: (returnval){ [ 1239.908909] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52641cdd-a39d-14bc-ec3a-999ee545370a" [ 1239.908909] env[62521]: _type = "HttpNfcLease" [ 1239.908909] env[62521]: } obtained for vApp import into resource pool (val){ [ 1239.908909] env[62521]: value = "resgroup-9" [ 1239.908909] env[62521]: _type = "ResourcePool" [ 1239.908909] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1239.909278] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the lease: (returnval){ [ 1239.909278] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52641cdd-a39d-14bc-ec3a-999ee545370a" [ 1239.909278] env[62521]: _type = "HttpNfcLease" [ 1239.909278] env[62521]: } to be ready. {{(pid=62521) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1239.915407] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1239.915407] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52641cdd-a39d-14bc-ec3a-999ee545370a" [ 1239.915407] env[62521]: _type = "HttpNfcLease" [ 1239.915407] env[62521]: } is initializing. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1240.024931] env[62521]: DEBUG oslo_concurrency.lockutils [req-15513a12-aa1e-478e-9bac-46c647311572 req-c583b1a5-0eed-48a4-b28f-7cb1139a2a89 service nova] Releasing lock "refresh_cache-89795492-b8e1-4b99-8451-3cc7a8ec3cb2" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1240.314622] env[62521]: DEBUG nova.network.neutron [-] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1240.419018] env[62521]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1240.419018] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52641cdd-a39d-14bc-ec3a-999ee545370a" [ 1240.419018] env[62521]: _type = "HttpNfcLease" [ 1240.419018] env[62521]: } is ready. {{(pid=62521) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1240.419018] env[62521]: DEBUG oslo_vmware.rw_handles [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1240.419018] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52641cdd-a39d-14bc-ec3a-999ee545370a" [ 1240.419018] env[62521]: _type = "HttpNfcLease" [ 1240.419018] env[62521]: }. {{(pid=62521) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1240.419018] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57b2bce-64fc-4b50-8f4b-747fc49b1501 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.425476] env[62521]: DEBUG oslo_vmware.rw_handles [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52109cc6-375e-9396-18b2-11d91ad9dae4/disk-0.vmdk from lease info. {{(pid=62521) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1240.425806] env[62521]: DEBUG oslo_vmware.rw_handles [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating HTTP connection to write to file with size = 31662080 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52109cc6-375e-9396-18b2-11d91ad9dae4/disk-0.vmdk. {{(pid=62521) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1240.490860] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3e16b497-6b8c-433f-bc4a-57af74fe14f7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.819906] env[62521]: INFO nova.compute.manager [-] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Took 1.24 seconds to deallocate network for instance. [ 1240.854902] env[62521]: DEBUG nova.compute.manager [req-0d6ebaa8-52bc-4fb2-b000-ed4c24645160 req-40c79a35-f1b8-47b2-a1cb-df4a4061e832 service nova] [instance: 144636f8-8650-4448-8fe5-aeec0ba6b69e] Received event network-vif-deleted-fe1d7288-4afc-44de-a9f7-4e98766784d4 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1241.327925] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1241.327925] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1241.327925] env[62521]: DEBUG nova.objects.instance [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lazy-loading 'resources' on Instance uuid 144636f8-8650-4448-8fe5-aeec0ba6b69e {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1241.573037] env[62521]: DEBUG oslo_vmware.rw_handles [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Completed reading data from the image iterator. {{(pid=62521) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1241.573318] env[62521]: DEBUG oslo_vmware.rw_handles [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52109cc6-375e-9396-18b2-11d91ad9dae4/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1241.574396] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e6a848-d33f-4fcf-90af-eac49eb80115 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.581567] env[62521]: DEBUG oslo_vmware.rw_handles [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52109cc6-375e-9396-18b2-11d91ad9dae4/disk-0.vmdk is in state: ready. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1241.581740] env[62521]: DEBUG oslo_vmware.rw_handles [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52109cc6-375e-9396-18b2-11d91ad9dae4/disk-0.vmdk. {{(pid=62521) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1241.581974] env[62521]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-45399c06-5644-4ed6-aa7c-5eea43c2c0bb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.792254] env[62521]: DEBUG oslo_vmware.rw_handles [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52109cc6-375e-9396-18b2-11d91ad9dae4/disk-0.vmdk. {{(pid=62521) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1241.792510] env[62521]: INFO nova.virt.vmwareapi.images [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Downloaded image file data 52ee4876-2464-4aaa-856d-801a073562b6 [ 1241.793348] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20975086-acfb-4bd3-b73c-9bfbc508e052 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.808069] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-40ba0a93-ac10-448e-8575-993826d0b302 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.838229] env[62521]: INFO nova.virt.vmwareapi.images [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] The imported VM was unregistered [ 1241.840649] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Caching image {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1241.840888] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Creating directory with path [datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6 {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1241.841205] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff958c9a-96fc-4f43-91f9-ad7311df8331 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.852282] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Created directory with path [datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6 {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1241.852282] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125/OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125.vmdk to [datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6/52ee4876-2464-4aaa-856d-801a073562b6.vmdk. {{(pid=62521) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1241.852454] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ba8614e4-301b-4a5b-972d-70681a324c30 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.858797] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1241.858797] env[62521]: value = "task-1319364" [ 1241.858797] env[62521]: _type = "Task" [ 1241.858797] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.868548] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319364, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.874521] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d33fd6-bb6d-4934-be31-ddadc38a3520 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.882633] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4b6d03-d122-4b31-9377-39f7a796ad75 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.911279] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b105e4e-e285-464d-9e38-6902732c6d8b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.918245] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b588bad4-5285-46f8-871b-5131cd114ab2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.930898] env[62521]: DEBUG nova.compute.provider_tree [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1242.369167] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319364, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.434526] env[62521]: DEBUG nova.scheduler.client.report [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1242.869948] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319364, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.939938] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.612s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1242.965189] env[62521]: INFO nova.scheduler.client.report [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Deleted allocations for instance 144636f8-8650-4448-8fe5-aeec0ba6b69e [ 1243.371301] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319364, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.473969] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f3b3a419-bf41-4897-9fbb-5f1b2fdaabd9 tempest-AttachInterfacesTestJSON-1439981337 tempest-AttachInterfacesTestJSON-1439981337-project-member] Lock "144636f8-8650-4448-8fe5-aeec0ba6b69e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.009s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.870717] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319364, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.372046] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319364, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.183553} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.372374] env[62521]: INFO nova.virt.vmwareapi.ds_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125/OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125.vmdk to [datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6/52ee4876-2464-4aaa-856d-801a073562b6.vmdk. [ 1244.372586] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Cleaning up location [datastore2] OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125 {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1244.372755] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_35b5299b-c62b-4c38-8016-437236cca125 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1244.373010] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aefcecfc-412b-4141-bec8-4c9988678436 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.379156] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1244.379156] env[62521]: value = "task-1319365" [ 1244.379156] env[62521]: _type = "Task" [ 1244.379156] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.387470] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319365, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.888504] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319365, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034373} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.888839] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1244.888930] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Releasing lock "[datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6/52ee4876-2464-4aaa-856d-801a073562b6.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1244.889199] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6/52ee4876-2464-4aaa-856d-801a073562b6.vmdk to [datastore2] 89795492-b8e1-4b99-8451-3cc7a8ec3cb2/89795492-b8e1-4b99-8451-3cc7a8ec3cb2.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1244.889451] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5086931-5772-4d65-be6e-f6134065d8d5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.896926] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1244.896926] env[62521]: value = "task-1319366" [ 1244.896926] env[62521]: _type = "Task" [ 1244.896926] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.902649] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319366, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.409581] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319366, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.906538] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319366, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.408645] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319366, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.908297] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319366, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.408939] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319366, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.275001} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.409233] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/52ee4876-2464-4aaa-856d-801a073562b6/52ee4876-2464-4aaa-856d-801a073562b6.vmdk to [datastore2] 89795492-b8e1-4b99-8451-3cc7a8ec3cb2/89795492-b8e1-4b99-8451-3cc7a8ec3cb2.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1247.410025] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056778d8-4beb-42c7-9e84-96b76001f03f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.433966] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 89795492-b8e1-4b99-8451-3cc7a8ec3cb2/89795492-b8e1-4b99-8451-3cc7a8ec3cb2.vmdk or device None with type streamOptimized {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1247.434268] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d670455e-315e-447b-b393-957120e2dca0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.456956] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1247.456956] env[62521]: value = "task-1319367" [ 1247.456956] env[62521]: _type = "Task" [ 1247.456956] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.466123] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319367, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.967356] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319367, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.467774] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319367, 'name': ReconfigVM_Task, 'duration_secs': 0.806644} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.468150] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 89795492-b8e1-4b99-8451-3cc7a8ec3cb2/89795492-b8e1-4b99-8451-3cc7a8ec3cb2.vmdk or device None with type streamOptimized {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1248.469027] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45a913b8-93e9-43c8-b7ff-074780a3e8e7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.475104] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1248.475104] env[62521]: value = "task-1319368" [ 1248.475104] env[62521]: _type = "Task" [ 1248.475104] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.486731] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319368, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.991450] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319368, 'name': Rename_Task, 'duration_secs': 0.137162} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.991450] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1248.991450] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff01e99d-79c1-4adb-8040-32ae1ebd3b9a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.997091] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1248.997091] env[62521]: value = "task-1319369" [ 1248.997091] env[62521]: _type = "Task" [ 1248.997091] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.004476] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319369, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.026810] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquiring lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.027052] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.505598] env[62521]: DEBUG oslo_vmware.api [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319369, 'name': PowerOnVM_Task, 'duration_secs': 0.423973} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.506053] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1249.529506] env[62521]: DEBUG nova.compute.manager [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1249.611486] env[62521]: DEBUG nova.compute.manager [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1249.612460] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a4588d-6c65-424b-8da5-2811b4cc4c71 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.050794] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.051104] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.052641] env[62521]: INFO nova.compute.claims [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1250.130774] env[62521]: DEBUG oslo_concurrency.lockutils [None req-7473bc1a-4b0f-4b5b-b437-46a23314ebb4 tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.974s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1250.745860] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.746097] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.746331] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.746557] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.746737] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1250.748744] env[62521]: INFO nova.compute.manager [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Terminating instance [ 1250.750537] env[62521]: DEBUG nova.compute.manager [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1250.750749] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1250.751667] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbcbf51a-8d2f-431d-8328-b169c0520346 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.759301] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1250.759528] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-458442ca-69e2-43fc-9bd7-6606dd8cbf30 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.766442] env[62521]: DEBUG oslo_vmware.api [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1250.766442] env[62521]: value = "task-1319370" [ 1250.766442] env[62521]: _type = "Task" [ 1250.766442] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.773722] env[62521]: DEBUG oslo_vmware.api [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319370, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.098822] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6903a63d-a1ef-4079-a745-6cc981a6a0bb {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.107691] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c05bdb-1d73-447a-bcb9-b1cf390fdafa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.136044] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ace440-ba0d-4ba9-b933-a920c2a62758 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.142774] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52da8039-6ed7-4313-b673-2a81a856b91a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.155421] env[62521]: DEBUG nova.compute.provider_tree [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1251.276319] env[62521]: DEBUG oslo_vmware.api [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319370, 'name': PowerOffVM_Task, 'duration_secs': 0.187449} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.276557] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1251.276740] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1251.276984] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed2de860-2308-4112-8ad5-1b15ece6620e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.362572] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1251.362809] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Deleting contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1251.362996] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleting the datastore file [datastore2] 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1251.363269] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c96f3a2-c44b-4d71-91a3-2ca6201907e1 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.370162] env[62521]: DEBUG oslo_vmware.api [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for the task: (returnval){ [ 1251.370162] env[62521]: value = "task-1319372" [ 1251.370162] env[62521]: _type = "Task" [ 1251.370162] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.377754] env[62521]: DEBUG oslo_vmware.api [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319372, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.658820] env[62521]: DEBUG nova.scheduler.client.report [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1251.880255] env[62521]: DEBUG oslo_vmware.api [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Task: {'id': task-1319372, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150315} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.880503] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1251.880691] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Deleted contents of the VM from datastore datastore2 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1251.880867] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1251.881058] env[62521]: INFO nova.compute.manager [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1251.881310] env[62521]: DEBUG oslo.service.loopingcall [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1251.881489] env[62521]: DEBUG nova.compute.manager [-] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1251.881581] env[62521]: DEBUG nova.network.neutron [-] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1252.164056] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.113s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.164622] env[62521]: DEBUG nova.compute.manager [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1252.415725] env[62521]: DEBUG nova.compute.manager [req-c9613d92-9052-4db7-a5b0-6ca6a4cbb5cf req-86d95c7e-ab54-4481-8e04-41f975b0d8e4 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Received event network-vif-deleted-c6359cbd-f300-4344-8a1b-bed3f0dccdf2 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1252.415725] env[62521]: INFO nova.compute.manager [req-c9613d92-9052-4db7-a5b0-6ca6a4cbb5cf req-86d95c7e-ab54-4481-8e04-41f975b0d8e4 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Neutron deleted interface c6359cbd-f300-4344-8a1b-bed3f0dccdf2; detaching it from the instance and deleting it from the info cache [ 1252.415725] env[62521]: DEBUG nova.network.neutron [req-c9613d92-9052-4db7-a5b0-6ca6a4cbb5cf req-86d95c7e-ab54-4481-8e04-41f975b0d8e4 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1252.669289] env[62521]: DEBUG nova.compute.utils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1252.670703] env[62521]: DEBUG nova.compute.manager [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1252.670893] env[62521]: DEBUG nova.network.neutron [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1252.708742] env[62521]: DEBUG nova.policy [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '475e1c3e9ea747c6a645d2c77a655a41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0882ccf7ebfc401ea5163801ee877df0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1252.890050] env[62521]: DEBUG nova.network.neutron [-] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1252.920046] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81d52b66-1fab-40ad-bfcf-77b00e8cf89a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.930696] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4be989a-920b-4bcc-b641-7a90ef2cd5c5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.943318] env[62521]: DEBUG nova.network.neutron [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Successfully created port: c86295b5-a3f2-471e-a04a-4d4b8ff02ab9 {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1252.956753] env[62521]: DEBUG nova.compute.manager [req-c9613d92-9052-4db7-a5b0-6ca6a4cbb5cf req-86d95c7e-ab54-4481-8e04-41f975b0d8e4 service nova] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Detach interface failed, port_id=c6359cbd-f300-4344-8a1b-bed3f0dccdf2, reason: Instance 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1253.174686] env[62521]: DEBUG nova.compute.manager [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1253.393485] env[62521]: INFO nova.compute.manager [-] [instance: 89795492-b8e1-4b99-8451-3cc7a8ec3cb2] Took 1.51 seconds to deallocate network for instance. [ 1253.900145] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.900437] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.900666] env[62521]: DEBUG nova.objects.instance [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lazy-loading 'resources' on Instance uuid 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1254.185209] env[62521]: DEBUG nova.compute.manager [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1254.215283] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1254.215550] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1254.215717] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1254.215905] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1254.216072] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1254.216229] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1254.216443] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1254.216667] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1254.216850] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1254.217028] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1254.217211] env[62521]: DEBUG nova.virt.hardware [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1254.218102] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684a6020-9de2-410c-9d22-3a2e5d12553c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.225730] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71058560-1253-48be-b4e9-efb9e16bd5fa {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.388211] env[62521]: DEBUG nova.network.neutron [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Successfully updated port: c86295b5-a3f2-471e-a04a-4d4b8ff02ab9 {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1254.439456] env[62521]: DEBUG nova.compute.manager [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Received event network-vif-plugged-c86295b5-a3f2-471e-a04a-4d4b8ff02ab9 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1254.439986] env[62521]: DEBUG oslo_concurrency.lockutils [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] Acquiring lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.440230] env[62521]: DEBUG oslo_concurrency.lockutils [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] Lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.440408] env[62521]: DEBUG oslo_concurrency.lockutils [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] Lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.440585] env[62521]: DEBUG nova.compute.manager [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] No waiting events found dispatching network-vif-plugged-c86295b5-a3f2-471e-a04a-4d4b8ff02ab9 {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1254.440758] env[62521]: WARNING nova.compute.manager [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Received unexpected event network-vif-plugged-c86295b5-a3f2-471e-a04a-4d4b8ff02ab9 for instance with vm_state building and task_state spawning. [ 1254.440925] env[62521]: DEBUG nova.compute.manager [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Received event network-changed-c86295b5-a3f2-471e-a04a-4d4b8ff02ab9 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1254.441240] env[62521]: DEBUG nova.compute.manager [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Refreshing instance network info cache due to event network-changed-c86295b5-a3f2-471e-a04a-4d4b8ff02ab9. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1254.441517] env[62521]: DEBUG oslo_concurrency.lockutils [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] Acquiring lock "refresh_cache-7eb4e8a9-2f79-48f5-9fec-8472de384ba6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1254.441666] env[62521]: DEBUG oslo_concurrency.lockutils [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] Acquired lock "refresh_cache-7eb4e8a9-2f79-48f5-9fec-8472de384ba6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.441833] env[62521]: DEBUG nova.network.neutron [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Refreshing network info cache for port c86295b5-a3f2-471e-a04a-4d4b8ff02ab9 {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1254.445838] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326ab6a6-47ca-4287-864e-70e9cfa61457 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.453992] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9edc2078-54e5-4742-927f-bea89a4c76d3 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.483347] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235b855e-eaa6-4a6b-b307-4e5c0c96916d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.490858] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735134b6-b386-4c02-8b62-4658e1bc0290 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.503531] env[62521]: DEBUG nova.compute.provider_tree [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1254.890417] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquiring lock "refresh_cache-7eb4e8a9-2f79-48f5-9fec-8472de384ba6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1254.971368] env[62521]: DEBUG nova.network.neutron [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1255.006245] env[62521]: DEBUG nova.scheduler.client.report [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1255.042571] env[62521]: DEBUG nova.network.neutron [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.512965] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.612s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.532008] env[62521]: INFO nova.scheduler.client.report [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Deleted allocations for instance 89795492-b8e1-4b99-8451-3cc7a8ec3cb2 [ 1255.545046] env[62521]: DEBUG oslo_concurrency.lockutils [req-9e17587c-ce0f-476a-a290-7e23220429b5 req-26777a27-096d-49f9-8ef4-bfe4e5e41a5f service nova] Releasing lock "refresh_cache-7eb4e8a9-2f79-48f5-9fec-8472de384ba6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1255.545385] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquired lock "refresh_cache-7eb4e8a9-2f79-48f5-9fec-8472de384ba6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.545543] env[62521]: DEBUG nova.network.neutron [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1256.039541] env[62521]: DEBUG oslo_concurrency.lockutils [None req-6fc2197b-4303-4096-8d40-5df1ef7355aa tempest-ServerActionsTestOtherB-673566239 tempest-ServerActionsTestOtherB-673566239-project-member] Lock "89795492-b8e1-4b99-8451-3cc7a8ec3cb2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.293s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.073385] env[62521]: DEBUG nova.network.neutron [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1256.208431] env[62521]: DEBUG nova.network.neutron [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Updating instance_info_cache with network_info: [{"id": "c86295b5-a3f2-471e-a04a-4d4b8ff02ab9", "address": "fa:16:3e:5a:6b:9b", "network": {"id": "da4988f7-f23a-4185-ae69-10da42171263", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1061203577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0882ccf7ebfc401ea5163801ee877df0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc86295b5-a3", "ovs_interfaceid": "c86295b5-a3f2-471e-a04a-4d4b8ff02ab9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.711648] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Releasing lock "refresh_cache-7eb4e8a9-2f79-48f5-9fec-8472de384ba6" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.712053] env[62521]: DEBUG nova.compute.manager [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Instance network_info: |[{"id": "c86295b5-a3f2-471e-a04a-4d4b8ff02ab9", "address": "fa:16:3e:5a:6b:9b", "network": {"id": "da4988f7-f23a-4185-ae69-10da42171263", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1061203577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0882ccf7ebfc401ea5163801ee877df0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc86295b5-a3", "ovs_interfaceid": "c86295b5-a3f2-471e-a04a-4d4b8ff02ab9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1256.712450] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:6b:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c86295b5-a3f2-471e-a04a-4d4b8ff02ab9', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1256.719902] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Creating folder: Project (0882ccf7ebfc401ea5163801ee877df0). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1256.720277] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f9cf50a2-9f0f-4066-b7fd-93643aa113bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.731011] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Created folder: Project (0882ccf7ebfc401ea5163801ee877df0) in parent group-v282025. [ 1256.731208] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Creating folder: Instances. Parent ref: group-v282211. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1256.731461] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b061e56d-8b5a-4ac5-aae9-25eecbd12bf7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.740500] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Created folder: Instances in parent group-v282211. [ 1256.740727] env[62521]: DEBUG oslo.service.loopingcall [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1256.740912] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1256.741123] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d168c068-4afd-4dd7-8eaa-5304ba1af4c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.758531] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1256.758531] env[62521]: value = "task-1319376" [ 1256.758531] env[62521]: _type = "Task" [ 1256.758531] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.765450] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319376, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.267868] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319376, 'name': CreateVM_Task, 'duration_secs': 0.266436} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.268050] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1257.268731] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1257.268912] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.269302] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1257.269565] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-633b4805-28dc-4a97-95ec-9b0eb08babb8 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.274069] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for the task: (returnval){ [ 1257.274069] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b04365-49bb-c277-466a-8e4f2d97708f" [ 1257.274069] env[62521]: _type = "Task" [ 1257.274069] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.281343] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b04365-49bb-c277-466a-8e4f2d97708f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.784047] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52b04365-49bb-c277-466a-8e4f2d97708f, 'name': SearchDatastore_Task, 'duration_secs': 0.009607} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.784365] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.784574] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1257.784818] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1257.784966] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.785170] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1257.785425] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d60d2cc-4816-42bd-9466-f677d7322b7b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.793351] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1257.793553] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1257.794310] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83bc2cc3-fee2-4c32-9746-d6b050c48fdd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.798999] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for the task: (returnval){ [ 1257.798999] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529eb0f2-77d7-fccf-65f2-2ab31c39916e" [ 1257.798999] env[62521]: _type = "Task" [ 1257.798999] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.806046] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529eb0f2-77d7-fccf-65f2-2ab31c39916e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.308694] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]529eb0f2-77d7-fccf-65f2-2ab31c39916e, 'name': SearchDatastore_Task, 'duration_secs': 0.007963} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.309426] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d2e385d-63d0-4a4b-9b47-da47806ad75d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.314668] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for the task: (returnval){ [ 1258.314668] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e05b7f-3ac9-49d1-6297-1b6ca5840400" [ 1258.314668] env[62521]: _type = "Task" [ 1258.314668] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.321714] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e05b7f-3ac9-49d1-6297-1b6ca5840400, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.824666] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e05b7f-3ac9-49d1-6297-1b6ca5840400, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.326206] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52e05b7f-3ac9-49d1-6297-1b6ca5840400, 'name': SearchDatastore_Task, 'duration_secs': 0.513972} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.326468] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1259.326726] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 7eb4e8a9-2f79-48f5-9fec-8472de384ba6/7eb4e8a9-2f79-48f5-9fec-8472de384ba6.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1259.326993] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a541610e-e210-47f0-9e30-7770023b5a01 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.333785] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for the task: (returnval){ [ 1259.333785] env[62521]: value = "task-1319377" [ 1259.333785] env[62521]: _type = "Task" [ 1259.333785] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.340716] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319377, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.847256] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319377, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475071} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.847256] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 7eb4e8a9-2f79-48f5-9fec-8472de384ba6/7eb4e8a9-2f79-48f5-9fec-8472de384ba6.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1259.847256] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1259.847256] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f9d1539-44d9-488c-8697-15a8a71ba3d9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.853060] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for the task: (returnval){ [ 1259.853060] env[62521]: value = "task-1319378" [ 1259.853060] env[62521]: _type = "Task" [ 1259.853060] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.859333] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319378, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.361232] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319378, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058509} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.361556] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1260.362360] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c9377b-ad3b-4791-bb46-154e5a8351a0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.384916] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 7eb4e8a9-2f79-48f5-9fec-8472de384ba6/7eb4e8a9-2f79-48f5-9fec-8472de384ba6.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1260.385040] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8e10163-f3e7-4125-aab3-7ef51eeca55c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.404799] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for the task: (returnval){ [ 1260.404799] env[62521]: value = "task-1319379" [ 1260.404799] env[62521]: _type = "Task" [ 1260.404799] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.412966] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319379, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.914821] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319379, 'name': ReconfigVM_Task, 'duration_secs': 0.306782} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.915144] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 7eb4e8a9-2f79-48f5-9fec-8472de384ba6/7eb4e8a9-2f79-48f5-9fec-8472de384ba6.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1260.915839] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a107e30-fa5c-4d30-8a92-327584fdf156 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.923066] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for the task: (returnval){ [ 1260.923066] env[62521]: value = "task-1319380" [ 1260.923066] env[62521]: _type = "Task" [ 1260.923066] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.931046] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319380, 'name': Rename_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.435258] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319380, 'name': Rename_Task, 'duration_secs': 0.135118} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.435258] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1261.435258] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46c46fad-3ef1-43d4-bca7-df3ddd87987a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.442155] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for the task: (returnval){ [ 1261.442155] env[62521]: value = "task-1319381" [ 1261.442155] env[62521]: _type = "Task" [ 1261.442155] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.449266] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319381, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.951549] env[62521]: DEBUG oslo_vmware.api [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319381, 'name': PowerOnVM_Task, 'duration_secs': 0.459236} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.955016] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1261.955016] env[62521]: INFO nova.compute.manager [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Took 7.77 seconds to spawn the instance on the hypervisor. [ 1261.955016] env[62521]: DEBUG nova.compute.manager [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1261.955016] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270e845e-03ff-47a8-be2a-53966a4df0da {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.475844] env[62521]: INFO nova.compute.manager [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Took 12.44 seconds to build instance. [ 1262.979322] env[62521]: DEBUG oslo_concurrency.lockutils [None req-74a4bab9-6ba9-47b8-9b25-b24313c2429d tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.952s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1263.541487] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquiring lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1263.541799] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1263.542072] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquiring lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1263.542220] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1263.542393] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1263.544681] env[62521]: INFO nova.compute.manager [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Terminating instance [ 1263.546504] env[62521]: DEBUG nova.compute.manager [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1263.546727] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1263.547546] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32642661-9e54-4077-9a15-fcd62153eab5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.555481] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1263.555719] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bab30450-46fd-4e80-ab6f-e2f466377ece {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.561371] env[62521]: DEBUG oslo_vmware.api [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for the task: (returnval){ [ 1263.561371] env[62521]: value = "task-1319382" [ 1263.561371] env[62521]: _type = "Task" [ 1263.561371] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.569226] env[62521]: DEBUG oslo_vmware.api [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.071376] env[62521]: DEBUG oslo_vmware.api [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319382, 'name': PowerOffVM_Task, 'duration_secs': 0.204946} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.072141] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1264.072141] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1264.072247] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2eed7a7-e9b6-4358-afd2-fd8c5bcfdf39 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.721791] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1265.723135] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1265.723135] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Deleting the datastore file [datastore1] 7eb4e8a9-2f79-48f5-9fec-8472de384ba6 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1265.723135] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-601f9a7f-88ea-41d5-b63c-0e9eda395e7c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.729006] env[62521]: DEBUG oslo_vmware.api [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for the task: (returnval){ [ 1265.729006] env[62521]: value = "task-1319384" [ 1265.729006] env[62521]: _type = "Task" [ 1265.729006] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.736697] env[62521]: DEBUG oslo_vmware.api [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319384, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.843557] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquiring lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1265.843842] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.239094] env[62521]: DEBUG oslo_vmware.api [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Task: {'id': task-1319384, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147782} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.239452] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1266.239694] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1266.239918] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1266.240162] env[62521]: INFO nova.compute.manager [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Took 2.69 seconds to destroy the instance on the hypervisor. [ 1266.240447] env[62521]: DEBUG oslo.service.loopingcall [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1266.240855] env[62521]: DEBUG nova.compute.manager [-] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1266.240855] env[62521]: DEBUG nova.network.neutron [-] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1266.346736] env[62521]: DEBUG nova.compute.manager [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Starting instance... {{(pid=62521) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1266.558601] env[62521]: DEBUG nova.compute.manager [req-dee9e36b-c30a-40c9-8941-019cc3216d42 req-0303ac59-1aef-4a90-8cf7-48d47034ef03 service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Received event network-vif-deleted-c86295b5-a3f2-471e-a04a-4d4b8ff02ab9 {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1266.558810] env[62521]: INFO nova.compute.manager [req-dee9e36b-c30a-40c9-8941-019cc3216d42 req-0303ac59-1aef-4a90-8cf7-48d47034ef03 service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Neutron deleted interface c86295b5-a3f2-471e-a04a-4d4b8ff02ab9; detaching it from the instance and deleting it from the info cache [ 1266.558988] env[62521]: DEBUG nova.network.neutron [req-dee9e36b-c30a-40c9-8941-019cc3216d42 req-0303ac59-1aef-4a90-8cf7-48d47034ef03 service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1266.871800] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.872082] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.873573] env[62521]: INFO nova.compute.claims [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1267.033613] env[62521]: DEBUG nova.network.neutron [-] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1267.066579] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-134871e5-4fff-4e37-b15b-4c32a1cfe6f5 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.075786] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6bee18-e68a-46dc-beb2-e7be0da2ee1a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.097683] env[62521]: DEBUG nova.compute.manager [req-dee9e36b-c30a-40c9-8941-019cc3216d42 req-0303ac59-1aef-4a90-8cf7-48d47034ef03 service nova] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Detach interface failed, port_id=c86295b5-a3f2-471e-a04a-4d4b8ff02ab9, reason: Instance 7eb4e8a9-2f79-48f5-9fec-8472de384ba6 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1267.536742] env[62521]: INFO nova.compute.manager [-] [instance: 7eb4e8a9-2f79-48f5-9fec-8472de384ba6] Took 1.30 seconds to deallocate network for instance. [ 1267.919972] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05420580-263e-4287-a930-73896a4f98bd {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.927273] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265a72bf-ed7e-4e95-89a4-4ead83b4b31d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.956412] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be03afc7-4d93-4313-bd9e-40cad30aad76 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.963443] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df56e46-981e-438a-95bd-b84629ebe17e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.976111] env[62521]: DEBUG nova.compute.provider_tree [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1268.042480] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.479406] env[62521]: DEBUG nova.scheduler.client.report [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1268.983975] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.112s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.984564] env[62521]: DEBUG nova.compute.manager [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Start building networks asynchronously for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1268.987199] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.945s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.987425] env[62521]: DEBUG nova.objects.instance [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lazy-loading 'resources' on Instance uuid 7eb4e8a9-2f79-48f5-9fec-8472de384ba6 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1269.491698] env[62521]: DEBUG nova.compute.utils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Using /dev/sd instead of None {{(pid=62521) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1269.496482] env[62521]: DEBUG nova.compute.manager [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Allocating IP information in the background. {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1269.496682] env[62521]: DEBUG nova.network.neutron [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] allocate_for_instance() {{(pid=62521) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1269.534023] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3a1191-e927-41a2-93f2-67a3d52ec587 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.541915] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c189904-67a4-4a3a-b112-8b983d1445b2 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.572675] env[62521]: DEBUG nova.policy [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5852cb59a5474224964e15246530f40b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24e1e3df761144bcbe73c46685985d71', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62521) authorize /opt/stack/nova/nova/policy.py:201}} [ 1269.574745] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e183be-b19d-4b2f-93c6-bacabdf5cd66 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.582689] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b3817e-7218-4a88-a367-a8d5e9bacb90 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.595755] env[62521]: DEBUG nova.compute.provider_tree [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1269.825329] env[62521]: DEBUG nova.network.neutron [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Successfully created port: 90779738-983f-4627-81ab-eb6531f5b46f {{(pid=62521) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1269.994743] env[62521]: DEBUG nova.compute.manager [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Start building block device mappings for instance. {{(pid=62521) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1270.098953] env[62521]: DEBUG nova.scheduler.client.report [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1270.605757] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1270.625693] env[62521]: INFO nova.scheduler.client.report [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Deleted allocations for instance 7eb4e8a9-2f79-48f5-9fec-8472de384ba6 [ 1271.004397] env[62521]: DEBUG nova.compute.manager [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Start spawning the instance on the hypervisor. {{(pid=62521) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1271.029926] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-20T00:28:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-20T00:28:25Z,direct_url=,disk_format='vmdk',id=58bd9a24-a0a4-4846-b299-475fa0f1d05d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c52a735a191b498eb13a1a29aec91182',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-20T00:28:26Z,virtual_size=,visibility=), allow threads: False {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1271.030208] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Flavor limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1271.030372] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Image limits 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1271.030563] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Flavor pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1271.030714] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Image pref 0:0:0 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1271.030868] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62521) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1271.031089] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1271.031256] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1271.031438] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Got 1 possible topologies {{(pid=62521) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1271.031637] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1271.031831] env[62521]: DEBUG nova.virt.hardware [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62521) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1271.032731] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a0f011-4221-462c-9e14-ae7be99cc3bf {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.040864] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70f82de-fbf8-4798-afce-1a1df5378bfc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.133538] env[62521]: DEBUG oslo_concurrency.lockutils [None req-26f6be46-7cd8-428a-b107-a6f29fc40895 tempest-ServerPasswordTestJSON-1858230895 tempest-ServerPasswordTestJSON-1858230895-project-member] Lock "7eb4e8a9-2f79-48f5-9fec-8472de384ba6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.592s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.481287] env[62521]: DEBUG nova.compute.manager [req-4658d77d-0d4a-4b62-b310-1b0a1b7777e8 req-4298d309-fbf9-4051-b2dc-985282fe10d8 service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Received event network-vif-plugged-90779738-983f-4627-81ab-eb6531f5b46f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1271.481524] env[62521]: DEBUG oslo_concurrency.lockutils [req-4658d77d-0d4a-4b62-b310-1b0a1b7777e8 req-4298d309-fbf9-4051-b2dc-985282fe10d8 service nova] Acquiring lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.481768] env[62521]: DEBUG oslo_concurrency.lockutils [req-4658d77d-0d4a-4b62-b310-1b0a1b7777e8 req-4298d309-fbf9-4051-b2dc-985282fe10d8 service nova] Lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1271.481946] env[62521]: DEBUG oslo_concurrency.lockutils [req-4658d77d-0d4a-4b62-b310-1b0a1b7777e8 req-4298d309-fbf9-4051-b2dc-985282fe10d8 service nova] Lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.483239] env[62521]: DEBUG nova.compute.manager [req-4658d77d-0d4a-4b62-b310-1b0a1b7777e8 req-4298d309-fbf9-4051-b2dc-985282fe10d8 service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] No waiting events found dispatching network-vif-plugged-90779738-983f-4627-81ab-eb6531f5b46f {{(pid=62521) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1271.483441] env[62521]: WARNING nova.compute.manager [req-4658d77d-0d4a-4b62-b310-1b0a1b7777e8 req-4298d309-fbf9-4051-b2dc-985282fe10d8 service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Received unexpected event network-vif-plugged-90779738-983f-4627-81ab-eb6531f5b46f for instance with vm_state building and task_state spawning. [ 1271.667107] env[62521]: DEBUG nova.network.neutron [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Successfully updated port: 90779738-983f-4627-81ab-eb6531f5b46f {{(pid=62521) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1272.173229] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquiring lock "refresh_cache-8b9e18da-c5bb-48ff-9c38-e738e56a1489" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1272.173229] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquired lock "refresh_cache-8b9e18da-c5bb-48ff-9c38-e738e56a1489" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1272.173229] env[62521]: DEBUG nova.network.neutron [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Building network info cache for instance {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1272.706031] env[62521]: DEBUG nova.network.neutron [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Instance cache missing network info. {{(pid=62521) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1272.840039] env[62521]: DEBUG nova.network.neutron [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Updating instance_info_cache with network_info: [{"id": "90779738-983f-4627-81ab-eb6531f5b46f", "address": "fa:16:3e:e7:52:1a", "network": {"id": "d1c70782-f9de-4744-90ce-b639b4a042eb", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-535125088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24e1e3df761144bcbe73c46685985d71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90779738-98", "ovs_interfaceid": "90779738-983f-4627-81ab-eb6531f5b46f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1273.316574] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1273.316888] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1273.317072] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62521) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1273.343040] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Releasing lock "refresh_cache-8b9e18da-c5bb-48ff-9c38-e738e56a1489" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1273.343131] env[62521]: DEBUG nova.compute.manager [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Instance network_info: |[{"id": "90779738-983f-4627-81ab-eb6531f5b46f", "address": "fa:16:3e:e7:52:1a", "network": {"id": "d1c70782-f9de-4744-90ce-b639b4a042eb", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-535125088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24e1e3df761144bcbe73c46685985d71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90779738-98", "ovs_interfaceid": "90779738-983f-4627-81ab-eb6531f5b46f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62521) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1273.343548] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:52:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c47e98ff-83cf-48d2-bf91-2931c7386b6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90779738-983f-4627-81ab-eb6531f5b46f', 'vif_model': 'vmxnet3'}] {{(pid=62521) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1273.350961] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Creating folder: Project (24e1e3df761144bcbe73c46685985d71). Parent ref: group-v282025. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1273.351238] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-033f2249-4f59-4861-ab2f-050ce2baf005 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.362275] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Created folder: Project (24e1e3df761144bcbe73c46685985d71) in parent group-v282025. [ 1273.362458] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Creating folder: Instances. Parent ref: group-v282214. {{(pid=62521) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1273.362678] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a028e51-9809-4069-9ff6-92248e135b37 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.371941] env[62521]: INFO nova.virt.vmwareapi.vm_util [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Created folder: Instances in parent group-v282214. [ 1273.372181] env[62521]: DEBUG oslo.service.loopingcall [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1273.372371] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Creating VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1273.372585] env[62521]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11861697-5432-44bb-b33f-c3c3c88ce08f {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.389671] env[62521]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1273.389671] env[62521]: value = "task-1319387" [ 1273.389671] env[62521]: _type = "Task" [ 1273.389671] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.396709] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319387, 'name': CreateVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.516450] env[62521]: DEBUG nova.compute.manager [req-1fd2d9b3-d61f-4856-897a-ff5b1e03b138 req-1061790d-d8c3-40d6-9c96-b782da8c60dd service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Received event network-changed-90779738-983f-4627-81ab-eb6531f5b46f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1273.516665] env[62521]: DEBUG nova.compute.manager [req-1fd2d9b3-d61f-4856-897a-ff5b1e03b138 req-1061790d-d8c3-40d6-9c96-b782da8c60dd service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Refreshing instance network info cache due to event network-changed-90779738-983f-4627-81ab-eb6531f5b46f. {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1273.516908] env[62521]: DEBUG oslo_concurrency.lockutils [req-1fd2d9b3-d61f-4856-897a-ff5b1e03b138 req-1061790d-d8c3-40d6-9c96-b782da8c60dd service nova] Acquiring lock "refresh_cache-8b9e18da-c5bb-48ff-9c38-e738e56a1489" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1273.517082] env[62521]: DEBUG oslo_concurrency.lockutils [req-1fd2d9b3-d61f-4856-897a-ff5b1e03b138 req-1061790d-d8c3-40d6-9c96-b782da8c60dd service nova] Acquired lock "refresh_cache-8b9e18da-c5bb-48ff-9c38-e738e56a1489" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1273.517256] env[62521]: DEBUG nova.network.neutron [req-1fd2d9b3-d61f-4856-897a-ff5b1e03b138 req-1061790d-d8c3-40d6-9c96-b782da8c60dd service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Refreshing network info cache for port 90779738-983f-4627-81ab-eb6531f5b46f {{(pid=62521) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1273.899087] env[62521]: DEBUG oslo_vmware.api [-] Task: {'id': task-1319387, 'name': CreateVM_Task, 'duration_secs': 0.321378} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.899274] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Created VM on the ESX host {{(pid=62521) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1273.899945] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1273.900131] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1273.900459] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1273.900709] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaafd99f-e38f-46b6-9699-df6caf344248 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.904945] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for the task: (returnval){ [ 1273.904945] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52843ca6-787b-e368-26f9-6fa575652471" [ 1273.904945] env[62521]: _type = "Task" [ 1273.904945] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.912121] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52843ca6-787b-e368-26f9-6fa575652471, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.227038] env[62521]: DEBUG nova.network.neutron [req-1fd2d9b3-d61f-4856-897a-ff5b1e03b138 req-1061790d-d8c3-40d6-9c96-b782da8c60dd service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Updated VIF entry in instance network info cache for port 90779738-983f-4627-81ab-eb6531f5b46f. {{(pid=62521) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1274.227441] env[62521]: DEBUG nova.network.neutron [req-1fd2d9b3-d61f-4856-897a-ff5b1e03b138 req-1061790d-d8c3-40d6-9c96-b782da8c60dd service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Updating instance_info_cache with network_info: [{"id": "90779738-983f-4627-81ab-eb6531f5b46f", "address": "fa:16:3e:e7:52:1a", "network": {"id": "d1c70782-f9de-4744-90ce-b639b4a042eb", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-535125088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24e1e3df761144bcbe73c46685985d71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90779738-98", "ovs_interfaceid": "90779738-983f-4627-81ab-eb6531f5b46f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1274.415560] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52843ca6-787b-e368-26f9-6fa575652471, 'name': SearchDatastore_Task, 'duration_secs': 0.0099} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.416264] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1274.416419] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Processing image 58bd9a24-a0a4-4846-b299-475fa0f1d05d {{(pid=62521) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1274.416667] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1274.416824] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquired lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.417016] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1274.417305] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97deee33-98e7-48f6-a9c7-541582e6c910 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.426722] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62521) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1274.426976] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62521) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1274.427764] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aac40128-fb98-4beb-8cb6-9a2fa6487185 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.433519] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for the task: (returnval){ [ 1274.433519] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5276d1c0-6021-c7da-4fcd-3ff09ba4ba57" [ 1274.433519] env[62521]: _type = "Task" [ 1274.433519] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.442065] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5276d1c0-6021-c7da-4fcd-3ff09ba4ba57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.730399] env[62521]: DEBUG oslo_concurrency.lockutils [req-1fd2d9b3-d61f-4856-897a-ff5b1e03b138 req-1061790d-d8c3-40d6-9c96-b782da8c60dd service nova] Releasing lock "refresh_cache-8b9e18da-c5bb-48ff-9c38-e738e56a1489" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1274.944752] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]5276d1c0-6021-c7da-4fcd-3ff09ba4ba57, 'name': SearchDatastore_Task, 'duration_secs': 0.010529} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.945603] env[62521]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6abb4aec-736f-48be-b1cb-fb9fff90fffc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.951276] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for the task: (returnval){ [ 1274.951276] env[62521]: value = "session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f2281b-0d1e-506b-92f1-769934f98476" [ 1274.951276] env[62521]: _type = "Task" [ 1274.951276] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.959383] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f2281b-0d1e-506b-92f1-769934f98476, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.462063] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': session[523a5516-6b7e-bff4-b24d-a602ce3fe5fa]52f2281b-0d1e-506b-92f1-769934f98476, 'name': SearchDatastore_Task, 'duration_secs': 0.016399} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.462482] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Releasing lock "[datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk" {{(pid=62521) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1275.462647] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 8b9e18da-c5bb-48ff-9c38-e738e56a1489/8b9e18da-c5bb-48ff-9c38-e738e56a1489.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1275.462960] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-239d49e5-9c1f-4b5c-b223-4277a1bf91c7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.469213] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for the task: (returnval){ [ 1275.469213] env[62521]: value = "task-1319388" [ 1275.469213] env[62521]: _type = "Task" [ 1275.469213] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.477806] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319388, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.978658] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319388, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496115} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.978930] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/58bd9a24-a0a4-4846-b299-475fa0f1d05d/58bd9a24-a0a4-4846-b299-475fa0f1d05d.vmdk to [datastore1] 8b9e18da-c5bb-48ff-9c38-e738e56a1489/8b9e18da-c5bb-48ff-9c38-e738e56a1489.vmdk {{(pid=62521) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1275.979165] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Extending root virtual disk to 1048576 {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1275.979441] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-670bbea7-c9b9-471a-8d26-8f0ef757e764 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.985094] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for the task: (returnval){ [ 1275.985094] env[62521]: value = "task-1319389" [ 1275.985094] env[62521]: _type = "Task" [ 1275.985094] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.992298] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319389, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.316904] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1276.317103] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Starting heal instance info cache {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1276.317226] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Rebuilding the list of instances to heal {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1276.495726] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319389, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.134968} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.496071] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Extended root virtual disk {{(pid=62521) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1276.496697] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bad9c07-f797-4e6f-b8b5-26271090e705 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.517436] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] 8b9e18da-c5bb-48ff-9c38-e738e56a1489/8b9e18da-c5bb-48ff-9c38-e738e56a1489.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1276.517655] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7153a531-48f7-4b84-abe3-3520259617cc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.536045] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for the task: (returnval){ [ 1276.536045] env[62521]: value = "task-1319390" [ 1276.536045] env[62521]: _type = "Task" [ 1276.536045] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.542994] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319390, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.820140] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Skipping network cache update for instance because it is Building. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1276.820316] env[62521]: DEBUG nova.compute.manager [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Didn't find any instances for network info cache update. {{(pid=62521) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1277.045860] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319390, 'name': ReconfigVM_Task, 'duration_secs': 0.275852} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.046147] env[62521]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Reconfigured VM instance instance-00000072 to attach disk [datastore1] 8b9e18da-c5bb-48ff-9c38-e738e56a1489/8b9e18da-c5bb-48ff-9c38-e738e56a1489.vmdk or device None with type sparse {{(pid=62521) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1277.046766] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da225beb-737c-487b-b312-a91ebd9f0878 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.052493] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for the task: (returnval){ [ 1277.052493] env[62521]: value = "task-1319391" [ 1277.052493] env[62521]: _type = "Task" [ 1277.052493] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.059530] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319391, 'name': Rename_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.315958] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.316216] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.316383] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager.update_available_resource {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.562242] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319391, 'name': Rename_Task, 'duration_secs': 0.135074} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.562613] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Powering on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1277.562724] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-722f92fd-107a-4d92-bb04-ca797e02df1d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.568971] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for the task: (returnval){ [ 1277.568971] env[62521]: value = "task-1319392" [ 1277.568971] env[62521]: _type = "Task" [ 1277.568971] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.576053] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.819667] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.819900] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.820036] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1277.820213] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62521) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1277.821147] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b019345a-a7d0-4f06-82e3-af7bd7e14b66 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.829223] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ce389c-3cc3-4c65-8abd-2bf181fa74d4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.844569] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56eda3f6-76ec-4998-8d97-cf25d5e6398b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.850896] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0460919-c238-46a6-a55e-696646296b5e {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.880871] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181007MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=62521) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1277.881061] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.881337] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1278.079024] env[62521]: DEBUG oslo_vmware.api [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319392, 'name': PowerOnVM_Task, 'duration_secs': 0.429708} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.079246] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Powered on the VM {{(pid=62521) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1278.079452] env[62521]: INFO nova.compute.manager [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Took 7.07 seconds to spawn the instance on the hypervisor. [ 1278.079646] env[62521]: DEBUG nova.compute.manager [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Checking state {{(pid=62521) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1278.080414] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd0e812-022b-4e27-8597-e85420e7038d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.596873] env[62521]: INFO nova.compute.manager [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Took 11.74 seconds to build instance. [ 1278.902810] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Instance 8b9e18da-c5bb-48ff-9c38-e738e56a1489 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62521) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1278.903032] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1278.903187] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62521) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1278.930465] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc39a49f-8573-4d56-b899-48c195dec7f0 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.938457] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29528ed-939b-4d77-94de-ae149a32999d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.969893] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4fc820-d2e7-423c-af97-12d8d78ceebc {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.977036] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9375d8ab-b561-48d5-8951-8f29bf34df9b {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.989677] env[62521]: DEBUG nova.compute.provider_tree [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1279.098944] env[62521]: DEBUG oslo_concurrency.lockutils [None req-2aba7533-dd17-4ba1-89e0-cc7a2f6a109c tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.255s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.336842] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquiring lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.337178] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1279.337408] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquiring lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.337603] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1279.337785] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.340073] env[62521]: INFO nova.compute.manager [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Terminating instance [ 1279.341995] env[62521]: DEBUG nova.compute.manager [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Start destroying the instance on the hypervisor. {{(pid=62521) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1279.342247] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Destroying instance {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1279.343124] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332e4ddf-99d8-41c7-9b60-99f992b52788 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.351092] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Powering off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1279.351327] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-591bf6ce-ebc6-46ed-9149-cb0dc8cb40a7 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.357260] env[62521]: DEBUG oslo_vmware.api [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for the task: (returnval){ [ 1279.357260] env[62521]: value = "task-1319393" [ 1279.357260] env[62521]: _type = "Task" [ 1279.357260] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.365125] env[62521]: DEBUG oslo_vmware.api [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.492655] env[62521]: DEBUG nova.scheduler.client.report [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1279.867082] env[62521]: DEBUG oslo_vmware.api [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319393, 'name': PowerOffVM_Task, 'duration_secs': 0.200855} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.867466] env[62521]: DEBUG nova.virt.vmwareapi.vm_util [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Powered off the VM {{(pid=62521) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1279.867567] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Unregistering the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1279.867803] env[62521]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5df301a-f914-42a6-adb4-94e52e2f154d {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.929422] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Unregistered the VM {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1279.929687] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Deleting contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1279.929881] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Deleting the datastore file [datastore1] 8b9e18da-c5bb-48ff-9c38-e738e56a1489 {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1279.930167] env[62521]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-76c7b93e-6dc1-4bcb-a38b-cf76c2de3ece {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.936724] env[62521]: DEBUG oslo_vmware.api [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for the task: (returnval){ [ 1279.936724] env[62521]: value = "task-1319395" [ 1279.936724] env[62521]: _type = "Task" [ 1279.936724] env[62521]: } to complete. {{(pid=62521) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.944257] env[62521]: DEBUG oslo_vmware.api [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319395, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.997279] env[62521]: DEBUG nova.compute.resource_tracker [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62521) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1279.997502] env[62521]: DEBUG oslo_concurrency.lockutils [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.116s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.446761] env[62521]: DEBUG oslo_vmware.api [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Task: {'id': task-1319395, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141838} completed successfully. {{(pid=62521) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.447142] env[62521]: DEBUG nova.virt.vmwareapi.ds_util [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Deleted the datastore file {{(pid=62521) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1280.447427] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Deleted contents of the VM from datastore datastore1 {{(pid=62521) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1280.447724] env[62521]: DEBUG nova.virt.vmwareapi.vmops [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Instance destroyed {{(pid=62521) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1280.447945] env[62521]: INFO nova.compute.manager [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1280.448202] env[62521]: DEBUG oslo.service.loopingcall [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62521) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1280.448469] env[62521]: DEBUG nova.compute.manager [-] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Deallocating network for instance {{(pid=62521) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1280.448576] env[62521]: DEBUG nova.network.neutron [-] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] deallocate_for_instance() {{(pid=62521) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1280.675603] env[62521]: DEBUG nova.compute.manager [req-2724ae42-5238-4227-aad9-632dfa180f7f req-49f85d05-913c-4ac6-8e78-cab6409ea240 service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Received event network-vif-deleted-90779738-983f-4627-81ab-eb6531f5b46f {{(pid=62521) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1280.675709] env[62521]: INFO nova.compute.manager [req-2724ae42-5238-4227-aad9-632dfa180f7f req-49f85d05-913c-4ac6-8e78-cab6409ea240 service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Neutron deleted interface 90779738-983f-4627-81ab-eb6531f5b46f; detaching it from the instance and deleting it from the info cache [ 1280.675931] env[62521]: DEBUG nova.network.neutron [req-2724ae42-5238-4227-aad9-632dfa180f7f req-49f85d05-913c-4ac6-8e78-cab6409ea240 service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1280.997986] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.998375] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.998375] env[62521]: DEBUG oslo_service.periodic_task [None req-f8de9e65-b04c-4d66-9339-49431a05185e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62521) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.152141] env[62521]: DEBUG nova.network.neutron [-] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Updating instance_info_cache with network_info: [] {{(pid=62521) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1281.178762] env[62521]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81ed80f0-6536-42bc-9e67-2d43ce8f2ac4 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.188087] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba0a7d3-acd2-4780-a46d-2c5993cc46b9 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.213989] env[62521]: DEBUG nova.compute.manager [req-2724ae42-5238-4227-aad9-632dfa180f7f req-49f85d05-913c-4ac6-8e78-cab6409ea240 service nova] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Detach interface failed, port_id=90779738-983f-4627-81ab-eb6531f5b46f, reason: Instance 8b9e18da-c5bb-48ff-9c38-e738e56a1489 could not be found. {{(pid=62521) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1281.655437] env[62521]: INFO nova.compute.manager [-] [instance: 8b9e18da-c5bb-48ff-9c38-e738e56a1489] Took 1.21 seconds to deallocate network for instance. [ 1282.161718] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.162039] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.162237] env[62521]: DEBUG nova.objects.instance [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lazy-loading 'resources' on Instance uuid 8b9e18da-c5bb-48ff-9c38-e738e56a1489 {{(pid=62521) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1282.697136] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0f78dc-27f9-4a5e-a7b5-c4b04f998d4c {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.704772] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d689a1-29fd-4dee-a943-cbba66155a70 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.733357] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6cfaab-d559-4eb1-ba39-94d65fb7ff9a {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.740150] env[62521]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c9ca48-9a9e-455d-bd29-1fed84c16415 {{(pid=62521) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.752804] env[62521]: DEBUG nova.compute.provider_tree [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Inventory has not changed in ProviderTree for provider: 3d21544b-bfc8-42d0-86ca-d323b5e2628f {{(pid=62521) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1283.256292] env[62521]: DEBUG nova.scheduler.client.report [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Inventory has not changed for provider 3d21544b-bfc8-42d0-86ca-d323b5e2628f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62521) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1283.761566] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1283.782936] env[62521]: INFO nova.scheduler.client.report [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Deleted allocations for instance 8b9e18da-c5bb-48ff-9c38-e738e56a1489 [ 1284.293155] env[62521]: DEBUG oslo_concurrency.lockutils [None req-91d33ed4-56a4-43f2-b444-c41acf60329f tempest-ServersNegativeTestMultiTenantJSON-1623962407 tempest-ServersNegativeTestMultiTenantJSON-1623962407-project-member] Lock "8b9e18da-c5bb-48ff-9c38-e738e56a1489" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.956s {{(pid=62521) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}